Three Models of RLHF Annotation: Extension, Evidence, and Authority

arXiv cs.CL / 4/29/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes Reinforcement Learning with Human Feedback (RLHF) by making explicit the usually-implicit normative role of human annotators’ judgments.
  • It proposes three conceptual models for how annotators influence system outputs: extension (extending designers’ judgments), evidence (providing independent facts), and authority (acting with independent standing to decide outputs).
  • The author argues that RLHF pipeline design should differ depending on which model best fits each annotation dimension, including solicitation, validation, and aggregation strategies.
  • By surveying key RLHF-related literature, the paper shows that many approaches implicitly combine these models and can fail when they conflate them.
  • The central recommendation is to decompose RLHF annotation into separable dimensions and build tailored, dimension-specific pipelines instead of using one unified annotation pipeline.

Abstract

Preference-based alignment methods, most prominently Reinforcement Learning with Human Feedback (RLHF), use the judgments of human annotators to shape large language model behaviour. However, the normative role of these judgments is rarely made explicit. I distinguish three conceptual models of that role. The first is extension: annotators extend the system designers' own judgments about what outputs should be. The second is evidence: annotators provide independent evidence about some facts, whether moral, social or otherwise. The third is authority: annotators have some independent authority (as representatives of the broader population) to determine system outputs. I argue that these models have implications for how RLHF pipelines should solicit, validate and aggregate annotations. I survey landmark papers in the literature on RLHF and related methods to illustrate how they implicitly draw on these models, describe failure modes that come from unintentionally or intentionally conflating them, and offer normative criteria for choosing among them. My central recommendation is that RLHF pipeline designers should decompose annotation into separable dimensions and tailor each pipeline to the model most appropriate for that dimension, rather than seeking a single unified pipeline.