Mitigating Reward Hacking in RLHF via Advantage Sign Robustness
arXiv cs.LG / 2026/4/6
📰 ニュースIdeas & Deep AnalysisModels & Research
要点
- The paper addresses reward hacking in RLHF, where reinforcement learning against a learned reward model can lead to true response quality plateauing or degrading.
- It argues that a key failure mode is “flipped advantage signs,” where an incorrect sign makes policy updates increase the likelihood of bad responses.
- By applying adversarial perturbations in the reward model parameter space, the authors derive a certified sign-preservation radius indicating the minimum perturbation needed to flip the advantage sign.
- They introduce Sign-Certified Policy Optimization (SignCert-PO), which down-weights policy-gradient contributions from non-robust (sign-unstable) completions.
- Experiments on TL;DR summarization and AlpacaFarm benchmarks show improved win rates over baselines and reduced reward hacking, with the method requiring only the RM parameters and on-policy completions at optimization time.




