Step-level Denoising-time Diffusion Alignment with Multiple Objectives

arXiv cs.LG / 4/17/2026

📰 NewsModels & Research

Key Points

  • The paper studies how to align diffusion models with human preferences when those preferences reflect multiple objectives rather than a single reward function.
  • It criticizes existing multi-objective methods for being either computationally expensive (multi-objective RL fine-tuning) or requiring reward access/gradients and introducing approximation error when merging objectives during denoising.
  • The authors introduce a step-level RL formulation to overcome the intractability of finding an optimal policy under KL regularization.
  • They propose MSDDA (Multi-objective Step-level Denoising-time Diffusion Alignment), a retraining-free framework that derives the optimal reverse denoising distribution in closed form using the mean and variance computed directly from single-objective base models.
  • The work proves the proposed denoising-time objective is exactly equivalent to step-level RL fine-tuning (no approximation error) and reports numerical results showing improved performance over prior denoising-time approaches.

Abstract

Reinforcement learning (RL) has emerged as a powerful tool for aligning diffusion models with human preferences, typically by optimizing a single reward function under a KL regularization constraint. In practice, however, human preferences are inherently pluralistic, and aligned models must balance multiple downstream objectives, such as aesthetic quality and text-image consistency. Existing multi-objective approaches either rely on costly multi-objective RL fine-tuning or on fusing separately aligned models at denoising time, but they generally require access to reward values (or their gradients) and/or introduce approximation error in the resulting denoising objectives. In this paper, we revisit the problem of RL fine-tuning for diffusion models and address the intractability of identifying the optimal policy by introducing a step-level RL formulation. Building on this, we further propose Multi-objective Step-level Denoising-time Diffusion Alignment (MSDDA), a retraining-free framework for aligning diffusion models with multiple objectives, obtaining the optimal reverse denoising distribution in closed form, with mean and variance expressed directly in terms of single-objective base models. We prove that this denoising-time objective is exactly equivalent to the step-level RL fine-tuning, introducing no approximation error. Moreover, we provide numerical results, which indicate our method outperforms existing denoising-time approaches.