Elucidating the SNR-t Bias of Diffusion Probabilistic Models

arXiv cs.CV / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies a Signal-to-Noise Ratio vs. timestep mismatch (SNR-t bias) in diffusion probabilistic models, where the SNR of the denoising sample no longer aligns with its timestep during inference.
  • It explains that while training tightly couples SNR to timestep, inference breaks this correspondence, causing error accumulation that degrades generation quality.
  • The authors support the claim with both empirical evidence and theoretical analysis, and introduce a differential correction approach.
  • By decomposing inputs into frequency components and applying correction per component—reflecting diffusion models’ tendency to recover low frequencies before high frequencies—the method substantially improves multiple diffusion model variants across datasets and resolutions with minimal compute cost.
  • The implementation is released at https://github.com/AMAP-ML/DCW.

Abstract

Diffusion Probabilistic Models have demonstrated remarkable performance across a wide range of generative tasks. However, we have observed that these models often suffer from a Signal-to-Noise Ratio-timestep (SNR-t) bias. This bias refers to the misalignment between the SNR of the denoising sample and its corresponding timestep during the inference phase. Specifically, during training, the SNR of a sample is strictly coupled with its timestep. However, this correspondence is disrupted during inference, leading to error accumulation and impairing the generation quality. We provide comprehensive empirical evidence and theoretical analysis to substantiate this phenomenon and propose a simple yet effective differential correction method to mitigate the SNR-t bias. Recognizing that diffusion models typically reconstruct low-frequency components before focusing on high-frequency details during the reverse denoising process, we decompose samples into various frequency components and apply differential correction to each component individually. Extensive experiments show that our approach significantly improves the generation quality of various diffusion models (IDDPM, ADM, DDIM, A-DPM, EA-DPM, EDM, PFGM++, and FLUX) on datasets of various resolutions with negligible computational overhead. The code is at https://github.com/AMAP-ML/DCW.