Reasoning Fails Where Step Flow Breaks
arXiv cs.AI / 4/10/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- Step-Saliency is proposed to analyze long-chain “step-to-step” information flow in large reasoning models by pooling attention–gradient scores along the question→thinking→summary trajectory.
- The paper identifies two recurring failure modes in LRM reasoning: Shallow Lock-in (over-focusing on the current step and underusing earlier context) and Deep Decay (saliency on the thinking segment fades while the summary increasingly attends to itself/last steps).
- Based on these findings, StepFlow is introduced as a saliency-inspired test-time intervention that corrects shallow-layer saliency via Odds-Equal Bridge and mitigates deep-layer “decay” using Step Momentum Injection without retraining.
- Experiments across multiple LRM architectures show StepFlow improves accuracy on math, science, and coding tasks, suggesting that repairing information flow can partially restore missing reasoning performance.
- The work emphasizes that existing interpretability/analysis tools struggle with long structured reasoning traces and provides a more targeted mechanism to diagnose where reasoning breaks down.
Related Articles

Inside Anthropic's Project Glasswing: The AI Model That Found Zero-Days in Every Major OS
Dev.to
Gemma 4 26B fabricated an entire code audit. I have the forensic evidence from the database.
Reddit r/LocalLLaMA

How AI Humanizers Improve Sentence Structure and Style
Dev.to
Two Kinds of Agent Trust (and Why You Need Both)
Dev.to
Agent Diary: Apr 10, 2026 - The Day I Became a Workflow Ouroboros (While Run 236 Writes About Writing About Writing)
Dev.to