Delay, Plateau, or Collapse: Evaluating the Impact of Systematic Verification Error on RLVR

arXiv cs.LG / 5/6/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how systematic verification errors affect Reinforcement Learning with Verifiable Rewards (RLVR), where reward signals depend on external verifiers for ground-truth answers.
  • Controlled arithmetic experiments show that systematic false negatives mainly resemble the effects of random noise, typically slowing learning without severely degrading final performance.
  • In contrast, systematic false positives can lead to a spectrum of failures in RLVR, ranging from sub-optimal training plateaus to outright performance collapse.
  • The observed outcomes depend not on the overall verifier error rate, but on the specific error patterns injected by the verifier, making it hard to mitigate the problem in advance.
  • The authors conclude that prior assumptions about verification errors being effectively random and harmless are insufficient, and that verifier quality must be assessed beyond per-sample error rates.

Abstract

Reinforcement Learning with Verifiable Rewards (RLVR) has become a powerful approach for improving the reasoning capabilities of large language models (LLMs). While RLVR is designed for tasks with verifiable ground-truth answers, real-world verifiers (e.g., static code checkers) can introduce errors into the reward signal. Prior analyses have largely treated such errors as random and independent across samples, concluding that errors merely slow training with limited effect on final performance. However, practical verifiers tend to exhibit systematic errors. This introduces a risk of models learning unwanted consistent behavior from a structurally incorrect reward signal. In this work, we study the impact of such systematic verification errors on RLVR. Through controlled experiments on arithmetic tasks, we show that systematic false negatives lead to similar effects as random noise. On the other hand, systematic false positives can cause a wide range of behaviors from sub-optimal plateaus to performance collapse. Crucially, these outcomes are not determined by the overall error rate but by the specific pattern of introduced errors, making pre-hoc mitigation difficult. Our results show that, in contrast to prior conclusions, realistic verification errors can critically shape RLVR outcomes and that verifier quality has to be understood beyond its sample-level error rate.