AI Navigate

Safe RLHF Beyond Expectation: Stochastic Dominance for Universal Spectral Risk Control

arXiv cs.LG / 3/12/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • RLHF safety currently relies on expected-cost constraints, which miss distributional tail risks, especially under heavy tails or rare events.
  • The work proposes Risk-sensitive Alignment via Dominance (RAD), replacing scalar expected-cost constraints with First-Order Stochastic Dominance constraints to control the entire cost distribution.
  • RAD is operationalized within an Optimal Transport framework using entropic regularization and Sinkhorn iterations to produce a differentiable, computationally efficient objective for end-to-end optimization.
  • The authors introduce quantile-weighted FSD constraints and show they universally control a broad class of Spectral Risk Measures, enabling tunable risk profiles with empirical improvements in harmlessness and robustness.

Abstract

Safe Reinforcement Learning from Human Feedback (RLHF) typically enforces safety through expected cost constraints, but the expectation captures only a single statistic of the cost distribution and fails to account for distributional uncertainty, particularly under heavy tails or rare catastrophic events. This limitation is problematic when robustness and risk sensitivity are critical. Stochastic dominance offers a principled alternative by comparing entire cost distributions rather than just their averages, enabling direct control over tail risks and potential out-of-distribution failures that expectation-based constraints may overlook. In this work, we propose Risk-sensitive Alignment via Dominance (RAD), a novel alignment framework that replaces scalar expected cost constraints with First-Order Stochastic Dominance (FSD) constraints. We operationalize this constraint by comparing the target policy's cost distribution to that of a reference policy within an Optimal Transport (OT) framework, using entropic regularization and Sinkhorn iterations to obtain a differentiable and computationally efficient objective for stable end-to-end optimization. Furthermore, we introduce quantile-weighted FSD constraints and show that weighted FSD universally controls a broad class of Spectral Risk Measures (SRMs), so that improvements under weighted dominance imply guaranteed improvements in the corresponding spectral risk. This provides a principled mechanism for tuning a model's risk profile via the quantile weighting function. Empirical results demonstrate that RAD improves harmlessness over baselines while remaining competitive in helpfulness, and exhibits greater robustness on out-of-distribution harmlessness evaluations.