AI Navigate

VC-Soup: Value-Consistency Guided Multi-Value Alignment for Large Language Models

arXiv cs.LG / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes VC-soup, a data filtering and parameter merging framework to address multi-value alignment in LLMs by emphasizing value consistency across different human values.
  • It introduces a value consistency metric based on the cosine similarity between the reward-gap vector of each preference pair and an all-ones vector, used to filter out low-consistency data.
  • Training on the remaining, value-consistent data yields policies that better preserve linear mode connectivity.
  • The approach linearly combines these value-specific policies and applies Pareto filtering across values to balance multi-value performance.
  • Experimental results and theoretical analysis indicate that VC-soup mitigates value conflicts and outperforms existing multi-value alignment methods.

Abstract

As large language models (LLMs) increasingly shape content generation, interaction, and decision-making across the Web, aligning them with human values has become a central objective in trustworthy AI. This challenge becomes even more pronounced when aligning multiple, potentially conflicting human values. Although recent approaches, such as reward reweighting, prompt-based supervised fine-tuning, and model merging, attempt to tackle multi-value alignment, they still face two major limitations: (1) training separate models for each value combination is prohibitively expensive; (2) value conflicts substantially degrade alignment performance. These limitations make it difficult to achieve favorable trade-offs across diverse human values. To address these challenges, we revisit multi-value alignment from the perspective of value consistency in data and propose VC-soup, a data filtering and parameter merging framework grounded in value-consistent learning. We first design a value consistency metric based on the cosine similarity between the reward-gap vector of each preference pair and an all-ones vector, which quantifies its cross-value coherence. We then filter out low-consistency preference pairs in each value dataset and train on the remaining data to obtain smooth, value-consistent policy models that better preserve linear mode connectivity. Finally, we linearly combine these policies and apply Pareto filtering across values to obtain solutions with balanced multi-value performance. Extensive experiments and theoretical analysis demonstrate that VC-soup effectively mitigates conflicts and consistently outperforms existing multi-value alignment methods.