Pref-CTRL: Preference Driven LLM Alignment using Representation Editing

arXiv cs.CL / 4/28/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • Pref-CTRL is a test-time LLM alignment approach that steers model outputs by making lightweight interventions on internal representations rather than fine-tuning the model weights.
  • The method addresses a gap in RE-Control by incorporating human preference structure, framing alignment as learning from preference judgments between candidate responses.
  • Pref-CTRL uses a multi-objective value function to better capture the objectives implied by preference data during representation editing.
  • Experiments on two benchmark datasets show Pref-CTRL outperforms RE-Control, with improved generalization on out-of-domain datasets.
  • The authors released source code on GitHub, enabling others to reproduce and build on the proposed framework.

Abstract

Test-time alignment methods offer a promising alternative to fine-tuning by steering the outputs of large language models (LLMs) at inference time with lightweight interventions on their internal representations. Recently, a prominent and effective approach, RE-Control (Kong et al., 2024), has proposed leveraging an external value function trained over the LLM's hidden states to guide generation via gradient-based editing. While effective, this method overlooks a key characteristic of alignment tasks, i.e. that they are typically formulated as learning from human preferences between candidate responses. To address this, in this paper we propose a novel preference-based training framework, Pref-CTRL, that uses a multi-objective value function to better reflect the structure of preference data. Our approach has outperformed RE-Control on two benchmark datasets and showed greater generalization on out-of-domain datasets. Our source code is available at https://github.com/UTS-nlPUG/pref-ctrl.