Rethinking the Comparison Unit in Sequence-Level Reinforcement Learning: An Equal-Length Paired Training Framework from Loss Correction to Sample Construction

arXiv cs.LG / 4/21/2026

📰 NewsModels & Research

Key Points

  • The paper argues that length-related issues in sequence-level relative reinforcement learning persist because training comparison units are not inherently comparable, not merely due to loss scaling or normalization bias.
  • It reframes the “length problem” as a comparison unit construction challenge and introduces a sample-construction-first training approach.
  • The proposed framework proactively generates equal-length, alignable, and comparable training segments, avoiding reliance on post-hoc corrections for unequal-length responses.
  • It presents EqLen, a method designed for group-relative comparison algorithms such as GRPO, GSPO, and RLOO, using techniques like dual-track synchronous generation, prefix inheritance, and segment masking to collect effective segments.
  • The overall goal is to enable more stable training by ensuring that the compared responses during generation are properly aligned and comparable.
  • Rethinking the Comparison Unit in Sequence-Level Reinforcement Learning: An Equal-Length Paired Training Framework from Loss Correction to Sample Construction

Abstract

This paper investigates the length problem in sequence-level relative reinforcement learning. We observe that, although existing methods partially alleviate length-related phenomena, a more fundamental issue remains insufficiently characterized: the comparison units used during training lack inherent comparability. Building on this observation, we propose a new perspective: the length problem should not be viewed merely as a loss-scaling or normalization bias, but rather as a \emph{comparison unit construction} problem. We further establish a sample-construction-based training framework that, instead of applying post-hoc corrections to unequal-length responses, proactively constructs equal-length, alignable, and comparable training segments during generation. Within this framework, we propose EqLen, a concrete method applicable to group-relative comparison algorithms such as GRPO, GSPO, and RLOO. Through dual-track synchronous generation, prefix inheritance, and segment masking, EqLen efficiently collects effective equal-length training segments and enables stable