ENGINEERING BLOG
Technical research and engineering insights from the team building the operating system for responsible AI operations.
188 articles · Published by MARIA OS
Eight papers that form the complete theory-to-operations stack: why organizational judgment needs an OS, structural design, stability laws, algorithm architecture, mission-constrained optimization, survival optimization, workforce transition, and agent lifecycle management.
Series Thesis
Company Intelligence explains why the OS exists. Structure defines responsibility. Stability laws prove when governance holds. Algorithms make it executable. Mission constraints keep optimization aligned. Survival theory determines evolutionary direction. White-collar transition shows who moves first. VITAL keeps the whole system alive.
00
Company Intelligence
Why organizational judgment needs an operating system, not just AI tools.
00
Company Intelligence
Why organizational judgment needs an operating system, not just AI tools.
01
Structural Design
How to decompose responsibility across human-agent boundaries.
02
Stability Laws
Mathematical conditions under which agentic governance holds or breaks.
03
Algorithm Stack
10 algorithms mapped to a 7-layer architecture for agentic organizations.
04
Mission Constraints
How to optimize agent goals without eroding organizational values.
05
Survival Optimization
Does evolutionary pressure reduce organizations to pure survival machines? The math of directed vs. undirected evolution.
06
Workforce Transition
Which white-collar workflows move first, and how fast the shift happens.
From memory and decision cards to strategic simulation, this is the architecture that turns AI Office from labor automation into an organization that learns
Most AI deployments improve local productivity but fail to compound into institutional intelligence. This article defines Company Intelligence as the closed loop of memory, decision, feedback, and governance, then explains how MARIA OS encodes that loop into company memory, executable decisions, agent performance systems, reflection pipelines, knowledge graphs, and strategic simulation.
AI Officeの価値は作業自動化ではなく、会社が記憶し、判断し、学習し、自己改善する閉ループを持てるかで決まる
多くのAI導入は局所的な生産性を改善しても、企業固有の知能には積み上がらない。本稿は、Company Intelligence を Memory・Decision・Feedback・Governance の閉ループとして定義し、MARIA OS がそれを Company Memory、Decision Card、Task Intelligence、Agent Performance、Knowledge Graph、Strategic Simulation へどう実装するかを解説する。
How agents recognize what they cannot do and trigger autonomous self-extension through formal gap analysis
Self-extending agents require a prerequisite that most architectures ignore: the ability to know what they do not know. This paper formalizes capability gap detection as a metacognitive layer that compares required capabilities against the agent's capability model, classifies detected gaps, prioritizes them by urgency and impact, and decides whether to synthesize, request, delegate, or escalate. We introduce the capability coverage metric, gap entropy measure, and multi-agent gap negotiation protocol. Experimental results show that agents with formal gap detection achieve 4.1x fewer silent failures and 2.8x faster self-extension compared to agents relying on runtime error detection.
形式的ギャップ分析を通じて、自分にできないことを認識し自律的な自己拡張をトリガーする方法
自己拡張型Agentには、ほとんどのアーキテクチャが無視する前提条件がある。自分に何ができないかを知る能力である。本論文はCapability Gap Detectionをメタ認知レイヤーとして形式化する。必要な能力をAgentの能力モデルと比較し、検出されたギャップを分類し、緊急度とインパクトで優先順位付けし、合成・要求・委任・エスカレーションの判断を下す。能力カバレッジメトリック、ギャップエントロピー測度、マルチAgent間ギャップ交渉プロトコルを導入する。
A formal architecture for encoding executive cognition into an auditable, drift-resistant persona layer that delegates judgment while preserving principal authority
Executive judgment is the highest-leverage bottleneck in any organization. Every strategic decision that waits for the CEO creates queue delay across the entire enterprise. Yet delegation through human hierarchies introduces information loss, preference distortion, and accountability diffusion. This paper presents the CEO Clone — not a chatbot that mimics speech patterns, but a computational decision interface that encodes the CEO's values, risk tolerance, decision patterns, and communication style into a formally verifiable persona layer. We model judgment delegation as a principal-agent problem with information asymmetry, introduce decision fidelity metrics with drift detection, and design calibration loops that maintain clone-principal alignment over time. The architecture operates within MARIA OS governance infrastructure, ensuring every delegated decision produces an immutable audit trail with full traceability to the encoded persona parameters that produced it.
経営者の認知を監査可能・ドリフト耐性のあるペルソナレイヤーとしてエンコードし、主体者の権限を保持しながら判断を委任する形式的アーキテクチャ
経営判断は、あらゆる組織において最もレバレッジの高いボトルネックである。CEOの判断を待つ全ての戦略的意思決定は、企業全体にキュー遅延を生む。しかし、人間の階層構造を通じた委任は、情報損失、選好歪曲、責任拡散を引き起こす。本論文では、CEOクローン——CEOの発話パターンを模倣するチャットボットではなく、CEOの価値観、リスク許容度、意思決定パターン、コミュニケーションスタイルを形式的に検証可能なペルソナレイヤーとしてエンコードする計算的意思決定インターフェース——を提示する。判断委任をプリンシパル・エージェント問題として情報の非対称性のもとでモデル化し、ドリフト検出を伴う意思決定忠実度メトリクスを導入し、クローンと主体者の整合性を長期にわたり維持するキャリブレーションループを設計する。本アーキテクチャはMARIA OSガバナンスインフラの下で運用され、全ての委任された意思決定が、それを生成したペルソナパラメータまで完全に追跡可能な不変の監査証跡を生成する。
経営認知を5次元意思決定空間 X = (L, D, G, I, R) として形式化し、判断重力・判断慣性・レイヤー整合の物理学で組織判断をスケールさせるCEO OSの完全設計論
判断はスケールしない。実行はスケールする。しかし、あらゆる組織は判断を人間の階層構造で積み重ねることでスケールさせようとし、各レイヤーで情報損失、選好歪曲、責任拡散を生み出す。CEO OSは組織判断を分類問題ではなく物理学の問題として扱う——重力、慣性、レイヤー、場を持つ力学系として。本論文は完全な意思決定力学の形式化を提示する:認知深度、ドメイン特化、判断重力、組織慣性、責任境界を捕捉する5軸意思決定空間 X = (L, D, G, I, R)。300問のベイズ推定型引き出しプロトコル、破滅的レイヤー不一致を防止するレイヤー整合アルゴリズム、モンテカルロシナリオ分析による反事実シミュレーションエンジンを導入する。本アーキテクチャは自己キャリブレーション型・ドリフト耐性の意思決定オペレーティングシステムを生成し、8.4倍の委任スループットと94.7%の判断忠実度を実現する。
Latent governance density, observable metacognitive coverage, and the stability bounds of self-governing enterprises
We formalize an agentic company as a graph-augmented constrained Markov decision process G_t = (A_t, E_t, S_t, Pi_t, R_t, D_t), distinguish latent governance density D_t from observable constrained-candidate coverage D_hat_t on router-generated Top-K actions, and define damping via kappa_t = kappa(D_hat_t). The exact local contraction condition is (1 - kappa_t) lambda_max(W_t) < 1, while the buffered operating envelope lambda_max(W_t) < 1 - kappa_t preserves adaptation headroom. Governance constraints thereby function as organizational metacognition: each constraint is a point where the system observes its own behavior. Planet-100 simulations validate that buffered role specialization emerges in the intermediate governance regime.
How self-improving routing uses recursive execution feedback to converge toward high-quality policies while preserving Lyapunov stability guarantees
Static action routing — where rules are configured once and applied uniformly — is inadequate for enterprise AI governance. Agent capabilities evolve, workloads shift, and routing quality depends on context that is only observed after execution. This paper introduces a recursive adaptation framework for MARIA OS action routing in which execution outcomes update routing parameters through a formal learning rule. We define θ_{t+1} = θ_t + η∇J(θ_t), where J(θ) is expected routing quality and gradients are estimated from outcome signals. We prove convergence under standard stochastic-approximation assumptions and establish Lyapunov stability guarantees, showing the adaptation process remains bounded while converging toward locally optimal routing policies. Thompson sampling provides principled exploration, and a multi-agent coordination protocol prevents oscillatory conflicts under concurrent adaptation. The quantitative figures in this article should be read as replay and simulation outputs over 14 operating contexts, not as audited production metrics of the current shipping router.
A formal analysis of how multi-agent teams calibrate collective confidence through structured interaction, showing why individual calibration is necessary but insufficient for team-level epistemic accuracy and how topology governs convergence
Individual calibration error measures how well one agent's stated confidence matches realized accuracy. In collaborative settings, however, a distinct phenomenon appears: collective calibration, where team-level confidence must track team-level accuracy. This paper defines collective calibration error as a metric that cannot be reduced to aggregated individual calibration, proves that individually well-calibrated agents can still form a poorly calibrated team under certain interaction topologies, and derives sufficient graph conditions for convergence. We validate the framework on MARIA OS deployments with 623 agents across 9 zones, showing a 41.7% reduction in collective calibration error via topology-aware reflection scheduling.
AGENT TEAMS FOR TECH BLOG
Every article passes through a 5-agent editorial pipeline. From research synthesis to technical review, quality assurance, and publication approval — each agent operates within its responsibility boundary.
Editor-in-Chief
ARIA-EDIT-01
Content strategy, publication approval, tone enforcement
G1.U1.P9.Z1.A1
Tech Lead Reviewer
ARIA-TECH-01
Technical accuracy, code correctness, architecture review
G1.U1.P9.Z1.A2
Writer Agent
ARIA-WRITE-01
Draft creation, research synthesis, narrative craft
G1.U1.P9.Z2.A1
Quality Assurance
ARIA-QA-01
Readability, consistency, fact-checking, style compliance
G1.U1.P9.Z2.A2
R&D Analyst
ARIA-RD-01
Benchmark data, research citations, competitive analysis
G1.U1.P9.Z3.A1
Distribution Agent
ARIA-DIST-01
Cross-platform publishing, EN→JA translation, draft management, posting schedule
G1.U1.P9.Z4.A1
Complete list of all 188 published articles. EN / JA bilingual index.
188 articles
All articles reviewed and approved by the MARIA OS Editorial Pipeline.
© 2026 MARIA OS. All rights reserved.