Generalized Distributional Alignment Games for Unbiased Answer-Level Fine-Tuning

arXiv stat.ML / 5/5/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a variational framework for Answer-Level Fine-Tuning (ALFT) using Distributional Alignment Games, but identifies that common implementations suffer from systematic estimation bias from logarithmic reward estimates on small batches.
  • It generalizes the game formulation to arbitrary Bregman divergences and provides provably unbiased estimators based on U-statistics for settings where polynomial-reward geometries are available.
  • For the canonical KL-divergence case where exact solutions are not feasible, the authors derive a globally robust minimax polynomial estimator with an optimal statistical error rate of Θ(1/K^2).
  • They further combine both threads into a Variance-Optimal Augmented Polynomial Optimization Program (AQP) estimator that reduces variance to improve bias and accelerates convergence, aiming for more efficient and stable training without extra online computation.

Abstract

The Distributional Alignment Game framework provides a powerful variational perspective on Answer-Level Fine-Tuning (ALFT). However, standard algorithms for these games rely on estimating logarithmic rewards from small batches, introducing a systematic bias due to Jensen's inequality that can destabilize training. In this paper, we systematically resolve this structural estimation bias. First, we generalize the alignment game to arbitrary Bregman divergences, showing that for a family of geometries inducing polynomial rewards, we can construct provably exact and unbiased estimators using U-statistics. Second, for the canonical KL divergence game where an exact solution is impossible, we derive a globally robust minimax polynomial estimator that is provably optimal, achieving the fundamental statistical error limit of \Theta(1/K^2), which we establish via the Ditzian-Totik theorem. Finally, we synthesize these two approaches to propose a novel Variance-Optimal Augmented Polynomial Optimization Program (AQP) Estimator, proving that by systematically reducing variance, our method achieves not only optimal bias but also provably accelerated game convergence, leading to more efficient and stable training with zero online computational overhead.