ENGINEERING BLOG
Technical research and engineering insights from the team building the operating system for responsible AI operations.
188 articles · Published by MARIA OS
Eight papers that form the complete theory-to-operations stack: why organizational judgment needs an OS, structural design, stability laws, algorithm architecture, mission-constrained optimization, survival optimization, workforce transition, and agent lifecycle management.
Series Thesis
Company Intelligence explains why the OS exists. Structure defines responsibility. Stability laws prove when governance holds. Algorithms make it executable. Mission constraints keep optimization aligned. Survival theory determines evolutionary direction. White-collar transition shows who moves first. VITAL keeps the whole system alive.
00
Company Intelligence
Why organizational judgment needs an operating system, not just AI tools.
00
Company Intelligence
Why organizational judgment needs an operating system, not just AI tools.
01
Structural Design
How to decompose responsibility across human-agent boundaries.
02
Stability Laws
Mathematical conditions under which agentic governance holds or breaks.
03
Algorithm Stack
10 algorithms mapped to a 7-layer architecture for agentic organizations.
04
Mission Constraints
How to optimize agent goals without eroding organizational values.
05
Survival Optimization
Does evolutionary pressure reduce organizations to pure survival machines? The math of directed vs. undirected evolution.
06
Workforce Transition
Which white-collar workflows move first, and how fast the shift happens.
Eliminating the command registry in favor of goal decomposition, plan generation, and dynamic tool synthesis
Traditional agent architectures bind agents to pre-defined command sets — fixed APIs, registered tools, and enumerated actions. This paper presents the MARIA OS command-less architecture, where agents receive goals rather than commands, decompose them into hierarchical plans, detect capability gaps, and synthesize whatever tools are needed for execution. We formalize the morphisms between Goal space G, Plan space P, and Tool space T, prove convergence of the tool space under recursive planning, and demonstrate that command-less agents achieve 3.2x higher task completion rates on novel problem classes compared to command-bound architectures.
コマンドレジストリを排除し、Goal分解・Plan生成・動的Tool合成によるAgent自律実行を実現する
従来のAgentアーキテクチャは事前定義されたコマンドセットに束縛される。本論文はMARIA OSのコマンドレスアーキテクチャを提示する。AgentはコマンドではなくGoalを受け取り、階層的Planに分解し、能力ギャップを検出し、必要なToolを動的に合成して実行する。Goal空間G、Plan空間P、Tool空間T間の射を形式化し、再帰的計画のもとでTool空間が収束することを証明する。
How agents recognize what they cannot do and trigger autonomous self-extension through formal gap analysis
Self-extending agents require a prerequisite that most architectures ignore: the ability to know what they do not know. This paper formalizes capability gap detection as a metacognitive layer that compares required capabilities against the agent's capability model, classifies detected gaps, prioritizes them by urgency and impact, and decides whether to synthesize, request, delegate, or escalate. We introduce the capability coverage metric, gap entropy measure, and multi-agent gap negotiation protocol. Experimental results show that agents with formal gap detection achieve 4.1x fewer silent failures and 2.8x faster self-extension compared to agents relying on runtime error detection.
形式的ギャップ分析を通じて、自分にできないことを認識し自律的な自己拡張をトリガーする方法
自己拡張型Agentには、ほとんどのアーキテクチャが無視する前提条件がある。自分に何ができないかを知る能力である。本論文はCapability Gap Detectionをメタ認知レイヤーとして形式化する。必要な能力をAgentの能力モデルと比較し、検出されたギャップを分類し、緊急度とインパクトで優先順位付けし、合成・要求・委任・エスカレーションの判断を下す。能力カバレッジメトリック、ギャップエントロピー測度、マルチAgent間ギャップ交渉プロトコルを導入する。
Beyond tool creation — a formal framework for bounded self-modification with stability guarantees and immutable audit trails
Agents that merely create new tools hit a ceiling. Real operational autonomy requires agents that can modify existing tools, rewrite commands, and restructure workflows based on performance feedback. We present a formal architecture for bounded self-modification with Lyapunov stability analysis, halting guarantees, and responsibility-gated audit trails.
ツール生成を超えて — 安定性保証と不変監査証跡を備えた有界自己修正の形式的フレームワーク
新しいツールを生成するだけのAgentには限界がある。真の運用自律性には、パフォーマンスフィードバックに基づいて既存のツール・コマンド・ワークフローを自ら書き換える能力が必要だ。本稿では、Lyapunov安定性解析・停止保証・責任ゲート付き監査証跡を備えた有界自己修正アーキテクチャSMASを提示する。
Agents as compilers — a formal framework mapping NL intent through intermediate representation to optimized, type-safe runtime tools
Tool-generating agents are ad-hoc code producers. We reframe tool synthesis as a compilation problem: natural language intent is parsed into an Intent AST, lowered to a Tool IR (intermediate representation), optimized through security hardening and dead code elimination passes, and emitted as type-safe executable code that hot-loads into the agent runtime. This paper presents the Agent Tool Compiler architecture with formal language theory foundations.
コンパイラとしてのAgent — NL意図を中間表現を経由して最適化された型安全なランタイムツールに変換する形式的フレームワーク
ツール生成Agentはアドホックなコード生産者である。本稿ではツール合成をコンパイル問題として再定義する。自然言語意図をIntent AST(意図の抽象構文木)に解析し、Tool IR(中間表現)に変換し、セキュリティ強化・デッドコード除去などの最適化パスを適用し、型安全な実行可能コードとしてエージェントランタイムにホットロードする。形式言語理論に基づくAgent Tool Compilerアーキテクチャを提示する。
Agents that recognize their own limitations and autonomously build the tools they need — within the safety boundaries of an operating system
Traditional AI agents are bounded by the tools humans provide. When an agent encounters a task outside its toolset, it halts and waits. This paper introduces the Self-Extending Agent Architecture (SEAA), where agents detect their own capability gaps, synthesize new tools through code generation, validate those tools in sandboxed environments, and register them into the OS runtime — all under human-governed safety constraints. We formalize the agent state model X_t = (C, T, M, R), derive the self-extension equation X_{t+1} = E_t ∘ G_t ∘ J_t(X_t), prove Capability Monotonicity under validation gates, and demonstrate the architecture within MARIA OS's hierarchical coordinate system.
Agentが自身の限界を検知し、コード生成でツールを合成し、サンドボックスで検証し、OSランタイムに登録する — すべてガバナンス制約の下で
従来のAIエージェントは、人間が提供したツールセットに束縛される。未対応タスクに遭遇すると停止し、人間の介入を待つ。本論文では、Self-Extending Agent Architecture(SEAA)を提案する。エージェントが自律的に能力ギャップを検出し、構造化コード生成でツールを合成し、サンドボックス環境で検証し、OSランタイムに登録するフレームワークである。エージェント状態モデル X_t = (C, T, M, R) を形式化し、自己拡張方程式 X_{t+1} = E_t ∘ G_t ∘ J_t(X_t) を導出し、検証ゲート下での能力単調性定理を証明する。MARIA OSの階層座標系における具体的な実装を示す。
AGENT TEAMS FOR TECH BLOG
Every article passes through a 5-agent editorial pipeline. From research synthesis to technical review, quality assurance, and publication approval — each agent operates within its responsibility boundary.
Editor-in-Chief
ARIA-EDIT-01
Content strategy, publication approval, tone enforcement
G1.U1.P9.Z1.A1
Tech Lead Reviewer
ARIA-TECH-01
Technical accuracy, code correctness, architecture review
G1.U1.P9.Z1.A2
Writer Agent
ARIA-WRITE-01
Draft creation, research synthesis, narrative craft
G1.U1.P9.Z2.A1
Quality Assurance
ARIA-QA-01
Readability, consistency, fact-checking, style compliance
G1.U1.P9.Z2.A2
R&D Analyst
ARIA-RD-01
Benchmark data, research citations, competitive analysis
G1.U1.P9.Z3.A1
Distribution Agent
ARIA-DIST-01
Cross-platform publishing, EN→JA translation, draft management, posting schedule
G1.U1.P9.Z4.A1
Complete list of all 188 published articles. EN / JA bilingual index.
188 articles
All articles reviewed and approved by the MARIA OS Editorial Pipeline.
© 2026 MARIA OS. All rights reserved.