PRISM: A Dual View of LLM Reasoning through Semantic Flow and Latent Computation

arXiv cs.CL / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • PRISM is introduced as a framework and diagnostic tool that jointly analyzes LLM reasoning across two levels: semantic token/step trajectories and latent internal computation across layers.
  • Using multiple reasoning models and benchmarks, the work identifies systematic patterns such as failed reasoning trajectories getting trapped in unproductive verification loops.
  • The analysis shows distinct divergence modes—e.g., overthinking versus premature commitment—that emerge differently once a candidate answer is reached.
  • PRISM demonstrates that prompting can reshape reasoning behavior not only in final accuracy, but also in how semantic transitions occur and how internal computational patterns evolve.
  • By modeling reasoning as structured processes, PRISM aims to make intermediate reasoning behaviors observable and diagnosable rather than relying solely on end-task accuracy.

Abstract

Large language models (LLMs) solve complex problems by generating multi-step reasoning traces. Yet these traces are typically analyzed from only one of two perspectives: the sequence of tokens across different reasoning steps in the generated text, or the hidden-state vectors across model layers within one step. We introduce PRISM (Probabilistic Reasoning Inspection through Semantic and Implicit Modeling), a framework and diagnostic tool for jointly analyzing both levels, providing a unified view of how reasoning evolves across steps and layers. Across multiple reasoning models and benchmarks, PRISM uncovers systematic patterns in the reasoning process, showing that failed trajectories are more likely to become trapped in unproductive verification loops and further diverge into distinct modes such as overthinking and premature commitment, which behave differently once a candidate answer is reached. It further reveals how prompting reshapes reasoning behavior beyond aggregate accuracy by altering both semantic transitions and internal computational patterns. By modeling reasoning trajectories as structured processes, PRISM makes these behaviors observable and analyzable rather than relying solely on final-task accuracy. Taken together, these insights position PRISM as a practical tool for analyzing and diagnosing reasoning processes in LLMs.