Cut Your Losses! Learning to Prune Paths Early for Efficient Parallel Reasoning

arXiv cs.CL / 4/20/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that parallel reasoning with Large Reasoning Models is often too costly because early mistakes lead to many unproductive reasoning paths.
  • It introduces the first systematic taxonomy for prefix-level path pruning, organizing approaches by signal source (internal vs. external) and whether the pruning is learnable or not.
  • Based on this taxonomy, it proposes STOP (Super TOken for Pruning), emphasizing the largely underexplored potential of learnable internal pruning methods.
  • Experiments across LRM sizes from 1.5B to 20B parameters show STOP is both more effective and more efficient than existing baselines.
  • The authors also demonstrate STOP’s scalability under different compute budgets (e.g., improving GPT-OSS-20B on AIME25 from 84% to nearly 90%) and provide empirical guidelines for real-world deployment, with code and models released online.

Abstract

Parallel reasoning enhances Large Reasoning Models (LRMs) but incurs prohibitive costs due to futile paths caused by early errors. To mitigate this, path pruning at the prefix level is essential, yet existing research remains fragmented without a standardized framework. In this work, we propose the first systematic taxonomy of path pruning, categorizing methods by their signal source (internal vs. external) and learnability (learnable vs. non-learnable). This classification reveals the unexplored potential of learnable internal methods, motivating our proposal of STOP (Super TOken for Pruning). Extensive evaluations across LRMs ranging from 1.5B to 20B parameters demonstrate that STOP achieves superior effectiveness and efficiency compared to existing baselines. Furthermore, we rigorously validate the scalability of STOP under varying compute budgets - for instance, boosting GPT-OSS-20B accuracy on AIME25 from 84% to nearly 90% under fixed compute budgets. Finally, we distill our findings into formalized empirical guidelines to facilitate optimal real-world deployment. Code, data and models are available at https://bijiaxihh.github.io/STOP