LLM Reasoning Is Latent, Not the Chain of Thought
arXiv cs.AI / 4/20/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that LLM “reasoning” should be studied as the formation of latent-state trajectories rather than as a faithful, observable chain-of-thought (CoT) on the surface.
- It explains that many conclusions—such as faithfulness, interpretability, reasoning benchmarks, and intervention-at-inference—depend on what researchers assume the primary unit of reasoning is.
- The authors disentangle three frequently confounded factors and formalize competing hypotheses: reasoning via latent trajectories (H1), reasoning via explicit surface CoT (H2), or reasoning gains driven mainly by generic serial compute (H0).
- By reorganizing prior empirical/mechanistic/survey evidence and adding compute-audited examples that separate surface traces from latent interventions and budget increases, the paper finds current evidence most strongly supports H1 as a default hypothesis.
- The paper recommends that the field adopt latent-state dynamics as the default object of study and evaluate reasoning using experimental designs that explicitly disentangle surface traces, latent states, and serial compute.
Related Articles
Which Version of Qwen 3.6 for M5 Pro 24g
Reddit r/LocalLLaMA

From Theory to Reality: Why Most AI Agent Projects Fail (And How Mine Did Too)
Dev.to

GPT-5.4-Cyber: OpenAI's Game-Changer for AI Security and Defensive AI
Dev.to

Building Digital Souls: The Brutal Reality of Creating AI That Understands You Like Nobody Else
Dev.to
Local LLM Beginner’s Guide (Mac - Apple Silicon)
Reddit r/artificial