FVD: Inference-Time Alignment of Diffusion Models via Fleming-Viot Resampling

arXiv cs.AI / 4/10/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Fleming-Viot Diffusion (FVD), an inference-time alignment method for diffusion samplers that targets diversity collapse and lineage collapse common in SMC-based approaches.
  • FVD replaces multinomial resampling with a Fleming-Viot-inspired birth-death mechanism, using independent reward-based survival decisions plus stochastic rebirth noise when rewards are only approximately available.
  • The approach aims to preserve broader trajectory support while still efficiently exploring reward-tilted distributions, and it does so without requiring value function approximation or costly rollouts.
  • The method is fully parallelizable and scales efficiently with inference compute, making it practical for larger sampling workloads.
  • Experiments report strong gains, including ~7% improvement on DrawBench for ImageReward, ~14–20% better FID on class-conditional tasks, and up to 66x faster than value-based approaches.

Abstract

We introduce Fleming-Viot Diffusion (FVD), an inference-time alignment method that resolves the diversity collapse commonly observed in Sequential Monte Carlo (SMC) based diffusion samplers. Existing SMC-based diffusion samplers often rely on multinomial resampling or closely related resampling schemes, which can still reduce diversity and lead to lineage collapse under strong selection pressure. Inspired by Fleming-Viot population dynamics, FVD replaces multinomial resampling with a specialized birth-death mechanism designed for diffusion alignment. To handle cases where rewards are only approximately available and naive rebirth would collapse deterministic trajectories, FVD integrates independent reward-based survival decisions with stochastic rebirth noise. This yields flexible population dynamics that preserve broader trajectory support while effectively exploring reward-tilted distributions, all without requiring value function approximation or costly rollouts. FVD is fully parallelizable and scales efficiently with inference compute. Empirically, it achieves substantial gains across settings: on DrawBench it outperforms prior methods by 7% in ImageReward, while on class-conditional tasks it improves FID by roughly 14-20% over strong baselines and is up to 66 times faster than value-based approaches.