Active Inference: A method for Phenotyping Agency in AI systems?

arXiv cs.AI / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that existing definitions of “agency” in AI are insufficient and proposes a minimal, inspectable framework based on intentionality, rationality, and explainability.
  • It instantiates the framework as a partially observable Markov decision process within a variational formulation where an agent’s action chain is driven by posterior beliefs, prior preferences, and minimization of expected free energy.
  • Using a T-maze benchmark, the authors show that empowerment (channel capacity between actions and anticipated observations) can operationally distinguish different agency phenotypes by changing the generative model structure.
  • The study concludes that when agents perform epistemic foraging to reduce ambiguity, effective governance should shift from relying mainly on external constraints toward modulating internal prior preferences.
  • Overall, the work provides a computationally grounded bridge between “agent phenotyping” methods and practical AI governance strategy.

Abstract

The proliferation of agentic artificial intelligence has outpaced the conceptual tools needed to characterize agency in computational systems. Prevailing definitions mainly rely on autonomy and goal-directedness. Here, we argue for a minimal notion open to principled inspection given three criteria: intentionality as action grounded in beliefs and desires, rationality as normatively coherent action entailed by a world model, and explainability as action causally traceable to internal states; we subsequently instantiate these as a partially observable Markov decision process under a variational framework wherein posterior beliefs, prior preferences, and the minimization of expected free energy jointly constitute an agentic action chain. Using a canonical T-maze paradigm, we evidence how empowerment, formulated as the channel capacity between actions and anticipated observations, serves as an operational metric that distinguishes zero-, intermediate-, and high-agency phenotypes through structural manipulations of the generative model. We conclude by arguing that as agents engage in epistemic foraging to resolve ambiguity, the governance controls that remain effective must shift systematically from external constraints to the internal modulation of prior preferences, offering a principled, variational bridge from computational phenotyping to AI governance strategy