AI Navigate

Runtime Governance for AI Agents: Policies on Paths

arXiv cs.AI / 3/18/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that AI agents' non-deterministic, path-dependent behavior cannot be fully governed at design time and that runtime governance must balance task completion with legal, data-breach, reputational, and other costs.
  • It formalizes compliance policies as deterministic functions that map agent identity, partial path, proposed next action, and organizational state to a policy violation probability.
  • It shows that prompt-level instructions and static access control are special cases of this framework, illustrating how these controls influence or constrain agent paths.
  • It argues that runtime evaluation of the path is the general approach needed for path-dependent policies, beyond static, design-time controls.
  • It presents a formal governance framework, concrete policy examples inspired by the AI Act, discusses a reference implementation, and identifies open problems including risk calibration and the limits of enforced compliance.

Abstract

AI agents -- systems that plan, reason, and act using large language models -- produce non-deterministic, path-dependent behavior that cannot be fully governed at design time, where with governed we mean striking the right balance between as high as possible successful task completion rate and the legal, data-breach, reputational and other costs associated with running agents. We argue that the execution path is the central object for effective runtime governance and formalize compliance policies as deterministic functions mapping agent identity, partial path, proposed next action, and organizational state to a policy violation probability. We show that prompt-level instructions (and "system prompts"), and static access control are special cases of this framework: the former shape the distribution over paths without actually evaluating them; the latter evaluates deterministic policies that ignore the path (i.e., these can only account for a specific subset of all possible paths). In our view, runtime evaluation is the general case, and it is necessary for any path-dependent policy. We develop the formal framework for analyzing AI agent governance, present concrete policy examples (inspired by the AI act), discuss a reference implementation, and identify open problems including risk calibration and the limits of enforced compliance.