Outcome Rewards Do Not Guarantee Verifiable or Causally Important Reasoning

arXiv cs.CL / 4/27/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces two evaluation metrics—Causal Importance of Reasoning (CIR) and Sufficiency of Reasoning (SR)—to test whether chain-of-thought reasoning learned via RLVR actually drives and explains model answers.
  • Experiments on Qwen2.5 and ReasoningGym show that although RLVR can improve task accuracy, it does not consistently increase CIR or SR, suggesting that reasoning chains may not be causally or evidentially central.
  • The authors find that applying a small amount of supervised fine-tuning (SFT) before RLVR can improve low CIR/SR when RLVR alone falls short.
  • They also demonstrate that CIR and SR can be improved without SFT by adding auxiliary CIR/SR rewards alongside outcome-based rewards, achieving RLVR-like accuracy with more causally important and sufficient reasoning.

Abstract

Reinforcement Learning from Verifiable Rewards (RLVR) on chain-of-thought reasoning has become a standard part of language model post-training recipes. A common assumption is that the reasoning chains trained through RLVR reliably represent how a model gets to its answer. In this paper, we develop two metrics for critically examining this assumption: Causal Importance of Reasoning (CIR), which measures the cumulative effect of reasoning tokens on the final answer, and Sufficiency of Reasoning (SR), which measures whether a verifier can arrive at an unambiguous answer based on the reasoning alone. Through experiments with the Qwen2.5 model series and ReasoningGym tasks, we find that: (1) while RLVR does improve task accuracy, it does not reliably improve CIR or SR, calling the role of reasoning in model performance into question; (2) a small amount of SFT before RLVR can be a remedy for low CIR and SR; and (3) CIR and SR can be improved even without SFT by applying auxiliary CIR/SR rewards on top of the outcome-based reward. This joint reward matches the accuracy of RLVR while also leading to causally important and sufficient reasoning. These results show that RLVR does not always lead models to rely on reasoning in the way that is commonly thought, but this issue can be remedied with simple modifications to the post-training procedure.