TriEx: A Game-based Tri-View Framework for Explaining Internal Reasoning in Multi-Agent LLMs

arXiv cs.CL / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • TriEx is a tri-view explainability framework for multi-agent LLM systems in interactive, partially observable environments.
  • It produces structured, evidence-anchored artifacts from three perspectives: the agent’s self-reasoning tied to actions, a time-evolving model of opponents’ beliefs, and third-person oracle audits grounded in environment reference signals.
  • By converting explanations from free-form narratives into checkable objects, TriEx enables comparisons of explanation faithfulness across time and viewpoints.
  • Experiments on imperfect-information strategic games show TriEx can analyze belief dynamics and evaluator reliability, exposing systematic gaps between what agents claim, what they believe, and what they actually do.
  • The paper argues that explainability is interaction-dependent and supports multi-view, evidence-grounded evaluation for LLM agents.

Abstract

Explainability for Large Language Model (LLM) agents is especially challenging in interactive, partially observable settings, where decisions depend on evolving beliefs and other agents. We present \textbf{TriEx}, a tri-view explainability framework that instruments sequential decision making with aligned artifacts: (i) structured first-person self-reasoning bound to an action, (ii) explicit second-person belief states about opponents updated over time, and (iii) third-person oracle audits grounded in environment-derived reference signals. This design turns explanations from free-form narratives into evidence-anchored objects that can be compared and checked across time and perspectives. Using imperfect-information strategic games as a controlled testbed, we show that TriEx enables scalable analysis of explanation faithfulness, belief dynamics, and evaluator reliability, revealing systematic mismatches between what agents say, what they believe, and what they do. Our results highlight explainability as an interaction-dependent property and motivate multi-view, evidence-grounded evaluation for LLM agents. Code is available at https://github.com/Einsam1819/TriEx.