AI Navigate

Enhancing Reinforcement Learning Fine-Tuning with an Online Refiner

arXiv cs.LG / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes dynamic constraints for reinforcement learning fine-tuning that intervene only when the model outputs degenerate responses, using an online refiner to generate a minimally corrected version while preserving content verbatim.
  • A reference model serves as the online refiner, producing a refined output that preserves verbatim content and fixes errors, which is then used to train the fine-tuned model with a supervised loss.
  • The mechanism automatically adjusts constraint strength based on output quality, strengthening or relaxing constraints as needed during training.
  • Experiments on dialogue and code generation show dynamic constraints outperform KL regularization and unconstrained baselines, achieving higher task rewards while maintaining training stability.

Abstract

Constraints are essential for stabilizing reinforcement learning fine-tuning (RFT) and preventing degenerate outputs, yet they inherently conflict with the optimization objective because stronger constraints limit the ability of a fine-tuned model to discover better solutions. We propose \textit{dynamic constraints} that resolve this tension by adapting to the evolving capabilities of the fine-tuned model based on the insight that constraints should only intervene when degenerate outputs occur. We implement this by using a reference model as an \textit{online refiner} that takes the response from the fine-tuned model and generates a minimally corrected version which preserves correct content verbatim while fixing errors. A supervised fine-tuning loss then trains the fine-tuned model to produce the refined output. This mechanism yields a constraint that automatically strengthens or relaxes based on output quality. Experiments on dialogue and code generation show that dynamic constraints outperform both KL regularization and unconstrained baselines, achieving substantially higher task rewards while maintaining training stability.