The Stepwise Informativeness Assumption: Why are Entropy Dynamics and Reasoning Correlated in LLMs?

arXiv cs.LG / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles a longstanding empirical puzzle in LLM research: why internal entropy dynamics (under the model’s predictive distribution) so strongly correlate with external correctness against ground-truth answers.
  • It proposes the Stepwise Informativeness Assumption (SIA), claiming that answer-relevant information accumulates in expectation along reasoning prefixes as generation progresses.
  • The authors argue SIA arises naturally from maximum-likelihood training on human reasoning traces and is reinforced by common fine-tuning and reinforcement-learning pipelines.
  • They derive testable observable signatures that connect conditional answer entropy patterns to likelihood of correctness.
  • Experiments across GSM8K, ARC, and SVAMP using multiple open-weight LLM families (e.g., Gemma-2, LLaMA-3.2, Qwen-2.5, DeepSeek, Olmo variants) show that training induces SIA and that correct reasoning traces exhibit characteristic conditional answer entropy behavior.

Abstract

Recent work uses entropy-based signals at multiple representation levels to study reasoning in large language models, but the field remains largely empirical. A central unresolved puzzle is why internal entropy dynamics, defined under the predictive distribution of a model, correlate so robustly with external correctness given by the ground-truth answer. In this paper, we argue that this correlation arises because autoregressive models reason correctly when they accumulate information about the true answer via answer-informative prefixes. We formalize this intuition via the Stepwise Informativeness Assumption (SIA), which states that reasoning prefixes accumulate answer-relevant information in expectation as generation progresses. We show that SIA naturally emerges from maximum-likelihood optimization on human reasoning traces and is reinforced by standard fine-tuning and reinforcement-learning pipelines. We then derive observable signatures of SIA linking conditional answer entropy dynamics to correctness. We empirically test SIA across multiple reasoning benchmarks (GSM8K, ARC, SVAMP) and a diverse set of open-weight LLMs (Gemma-2, LLaMA-3.2, Qwen-2.5, DeepSeek and Olmo variants), showing that training induces it and that correct traces exhibit characteristic conditional answer entropy patterns.