Why Do LLMs Struggle in Strategic Play? Broken Links Between Observations, Beliefs, and Actions

arXiv cs.AI / 5/4/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper examines why LLMs can struggle at strategic decision-making under incomplete information, focusing on two internal mechanism gaps uncovered via experiments.
  • It finds an “observation–belief gap”: LLMs form internal beliefs about hidden game states that are more accurate than their spoken explanations, but those beliefs become brittle and degrade with multi-hop reasoning.
  • It reports “bias and coherence” issues in those beliefs, including primacy and recency effects as well as drift away from Bayesian coherence during longer interactions.
  • It identifies a “belief–action gap”: LLMs’ conversion of internal beliefs into actions appears weaker than the influence of beliefs provided in prompts, and belief-conditioning does not reliably improve payoffs.
  • The authors conclude that analyzing LLM internal processes reveals systematic vulnerabilities, suggesting caution when deploying LLMs in strategic domains without strong guardrails.

Abstract

Large language models (LLMs) are increasingly tasked with strategic decision-making under incomplete information, such as in negotiation and policymaking. While LLMs can excel at many such tasks, they also fail in ways that are poorly understood. We shed light on these failures by uncovering two fundamental gaps in the internal mechanisms underlying the decision-making of LLMs in incomplete-information games, supported by experiments with open-weight models Llama 3.1, Qwen3, and gpt-oss. First, an observation-belief gap: LLMs encode internal beliefs about latent game states that are substantially more accurate than their own verbal reports, yet these beliefs are brittle. In particular, the belief accuracy degrades with multi-hop reasoning, exhibits primacy and recency biases, and drifts away from Bayesian coherence over extended interactions. Second, a belief-action gap: The implicit conversion of internal beliefs into actions is weaker than that of the beliefs externalized in the prompt, yet neither belief-conditioning consistently achieves higher game payoffs. These results show how analyzing LLMs' internal processes can expose systematic vulnerabilities that warrant caution before deploying LLMs in strategic domains without robust guardrails.