AI Navigate

LLMs Faithfully and Iteratively Compute Answers During CoT: A Systematic Analysis With Multi-step Arithmetics

arXiv cs.CL / 3/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study analyzes how LLMs perform chain-of-thought reasoning and whether the final answer is determined before or during the CoT process, with a focus on faithfulness.
  • Experiments on controlled arithmetic tasks show that LLMs compute sub-answers while generating the reasoning chain, rather than deriving the final answer after input, indicating that internal computation is reflected in the chain.
  • The results indicate that chain-of-thought explanations can faithfully reflect the model's internal computations, challenging the view that CoT is just post-hoc rationalization.
  • The findings have implications for prompt design, evaluation of CoT-based systems, and how practitioners interpret model reasoning in real-world AI applications.

Abstract

This study investigates the internal information flow of large language models (LLMs) while performing chain-of-thought (CoT) style reasoning. Specifically, with a particular interest in the faithfulness of the CoT explanation to LLMs' final answer, we explore (i) when the LLMs' answer is (pre)determined, especially before the CoT begins or after, and (ii) how strongly the information from CoT specifically has a causal effect on the final answer. Our experiments with controlled arithmetic tasks reveal a systematic internal reasoning mechanism of LLMs. They have not derived an answer at the moment when input was fed into the model. Instead, they compute (sub-)answers while generating the reasoning chain on the fly. Therefore, the generated reasoning chains can be regarded as faithful reflections of the model's internal computation.