ConsistRM: Improving Generative Reward Models via Consistency-Aware Self-Training

arXiv cs.CL / 4/10/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ConsistRM, a self-training framework for generative reward models (GRMs) that aims to align LLMs with human preferences without requiring costly human-annotated reward data.
  • It proposes a Consistency-Aware Answer Reward to generate pseudo-labels that are reliable and maintain temporal consistency, improving stability of GRM training and optimization.
  • It also adds a Consistency-Aware Critique Reward that evaluates semantic consistency across multiple critiques and assigns fine-grained, differentiated rewards to reduce weaknesses seen in prior self-training methods.
  • Experiments across five benchmark datasets and four base models show ConsistRM outperforms vanilla reinforcement fine-tuning (RFT) by an average of 1.5%, while analysis indicates better output consistency and reduced position bias from input order.

Abstract

Generative reward models (GRMs) have emerged as a promising approach for aligning Large Language Models (LLMs) with human preferences by offering greater representational capacity and flexibility than traditional scalar reward models. However, GRMs face two major challenges: reliance on costly human-annotated data restricts scalability, and self-training approaches often suffer from instability and vulnerability to reward hacking. To address these issues, we propose ConsistRM, a self-training framework that enables effective and stable GRM training without human annotations. ConsistRM incorporates the Consistency-Aware Answer Reward, which produces reliable pseudo-labels with temporal consistency, thereby providing more stable model optimization. Moreover, the Consistency-Aware Critique Reward is introduced to assess semantic consistency across multiple critiques and allocates fine-grained and differentiated rewards. Experiments on five benchmark datasets across four base models demonstrate that ConsistRM outperforms vanilla Reinforcement Fine-Tuning (RFT) by an average of 1.5%. Further analysis shows that ConsistRM enhances output consistency and mitigates position bias caused by input order, highlighting the effectiveness of consistency-aware rewards in improving GRMs.