Design Conditions for Intra-Group Learning of Sequence-Level Rewards: Token Gradient Cancellation

arXiv cs.AI / 4/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how sparse termination rewards in intra-group RL fine-tuning of reasoning models can degrade long-horizon training through learning tax, solution probability drift, and entropy collapse.
  • It derives a token-level credit-assignment design condition requiring intra-group objectives to preserve gradient exchangeability across token updates so that weak-credit/high-frequency tokens undergo effective gradient cancellation.
  • The authors argue that two widely used mechanisms break this exchangeability, making non-cancellation a structural outcome in typical training setups.
  • They propose minimal intra-group objective transformations to restore or approximate the cancellation structure in the shared token space.
  • Experiments indicate these transformations stabilize training dynamics, improve sample efficiency, and increase final model performance, supporting the proposed design principle.

Abstract

In sparse termination rewards, intra-group comparisons have become the dominant paradigm for fine-tuning reasoning models via reinforcement learning. However, long-term training often leads to issues like ineffective update accumulation (learning tax), solution probability drift, and entropy collapse. This paper presents a necessary condition for algorithm design from a token-level credit assignment perspective: to prevent reward-irrelevant drift, intra-group objectives must maintain gradient exchangeability across token updates, enabling gradient cancellation on weak-credit/high-frequency tokens. We show that two common mechanisms disrupting exchangeability make "non-cancellation" a structural norm. Based on this, we propose minimal intra-group transformations to restore or approximate the cancellation structure in the shared token space. Experimental results demonstrate that these transformations stabilize training, improve sample efficiency, and enhance final performance, validating the value of this design condition.