NeuReasoner: Towards Explainable, Controllable, and Unified Reasoning via Mixture-of-Neurons
arXiv cs.CL / 4/6/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that large reasoning models suffer recurring failure modes at three levels—within-step errors, inter-step oscillation/stagnation, and instance-level maladaptive overthinking—yet existing work treats these issues in isolation.
- It presents a white-box analysis that uses a Mixture-of-Neurons (MoN) perspective to identify key neurons and their fluctuation patterns tied to specific failure types.
- Based on these findings, the authors propose NeuReasoner, a unified framework aimed at making reasoning explainable and controllable via MoN-driven mechanisms.
- NeuReasoner combines lightweight MLPs for failure detection with a special token-triggered self-correction learned through supervised fine-tuning (SFT), inserting tokens during inference to activate remedial behaviors.
- Experiments across six benchmarks and six backbone model sizes (8B–70B) show up to 27.0% performance gains and 19.6%–63.3% reductions in token consumption versus nine baselines.




