SPS: Steering Probability Squeezing for Better Exploration in Reinforcement Learning for Large Language Models

arXiv cs.CL / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that RL fine-tuning for reasoning models often over-optimizes single-sample success (Pass@1) while limiting exploration of diverse reasoning paths needed for better multi-sample performance (Pass@k).
  • It attributes this to a “probability mass squeezing” effect, where probability becomes overly concentrated on a small set of high-reward trajectories, reducing genuine trajectory diversity.
  • To counter the squeezing, the authors propose Steering Probability Squeezing (SPS), which alternates standard RL with inverse reinforcement learning (IRL) and treats on-policy rollouts as demonstrations to reshape the trajectory distribution.
  • Experiments on five reasoning benchmarks show SPS improves exploration and yields higher Pass@k, and the work also analyzes RL learning dynamics to estimate an empirical upper bound on achievable Pass@k.
  • Overall, the results suggest that alternating RL and IRL can extend the intrinsic exploration capability of RL-trained large language model reasoning systems without relying on external supervision.

Abstract

Reinforcement learning (RL) has emerged as a promising paradigm for training reasoning-oriented models by leveraging rule-based reward signals. However, RL training typically tends to improve single-sample success rates (i.e., Pass@1) while offering limited exploration of diverse reasoning trajectories, which is crucial for multi-sample performance (i.e., Pass@k). Our preliminary analysis reveals that this limitation stems from a fundamental squeezing effect, whereby probability mass is excessively concentrated on a narrow subset of high-reward trajectories, restricting genuine exploration and constraining attainable performance under RL training. To address this issue, in this work, we propose Steering Probability Squeezing (SPS), a training paradigm that interleaves conventional RL with inverse reinforcement learning (IRL). SPS treats on-policy rollouts as demonstrations and employs IRL to explicitly reshape the induced trajectory distribution, thereby enhancing exploration without introducing external supervision. Experiments on five commonly used reasoning benchmarks demonstrate that SPS can enable better exploration and improve Pass@k. Beyond algorithmic contributions, we provide an analysis of RL learning dynamics and identify an empirical upper bound on Pass@k, shedding light on intrinsic exploration limits in RL-based reasoning models. Our findings suggest that alternating between RL and IRL offers an effective pathway toward extending the exploration capacity of reasoning-oriented large language models.