Agentic AI systems - capable of goal interpretation, world modeling, planning, tool use, long-horizon operation, and autonomous coordination - introduce distinct control failures not addressed by existing safety frameworks. We identify six agentic governance failures tied to these capabilities and show how they erode meaningful human control in military settings. We propose the Agentic Military AI Governance Framework (AMAGF), a measurable architecture structured around three pillars: Preventive Governance (reducing failure likelihood), Detective Governance (real-time detection of control degradation), and Corrective Governance (restoring or safely degrading operations). Its core mechanism, the Control Quality Score (CQS), is a composite real-time metric quantifying human control and enabling graduated responses as control weakens. For each failure type, we define concrete mechanisms, assign responsibilities across five institutional actors, and formalize evaluation metrics. A worked operational scenario illustrates implementation, and we situate the framework within established agent safety literature. We argue that governance must move from a binary conception of control to a continuous model in which control quality is actively measured and managed throughout the operational lifecycle.
翻译:智能体AI系统——具备目标解读、世界建模、规划、工具使用、长时程操作与自主协同能力——会引发现有安全框架未能涵盖的特定控制失效问题。我们识别出与这些能力相关的六类智能体治理失效模式,并阐明它们如何侵蚀军事场景中有意义的人类控制。本文提出智能体军事AI治理框架(AMAGF),该可量化架构围绕三大支柱构建:预防性治理(降低失效概率)、侦测性治理(实时识别控制退化)与矫正性治理(恢复运行或实现安全降级)。其核心机制——控制质量评分(CQS)——是一个复合实时度量指标,可量化人类控制水平,并在控制弱化时启动分级响应机制。针对每类失效模式,我们定义了具体机制,划分了五类制度主体的责任,并形式化了评估指标。通过典型作战场景推演说明实施流程,并将本框架置于现有智能体安全研究谱系中。我们认为治理范式必须从二元控制观念转向连续控制模型,即在作战全生命周期中持续度量并动态管理控制质量。