A Probabilistic Formulation of Offset Noise in Diffusion Models

arXiv stat.ML / 4/10/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a known weakness of diffusion models when generating data with extreme brightness values and notes that while offset noise helps empirically, its theoretical foundation is still limited.
  • It proposes a new diffusion-model formulation that injects additional noise through a rigorous probabilistic framework by modifying both the forward and reverse diffusion processes.
  • The method allows diffusion of inputs into Gaussian distributions with arbitrary mean structures and derives a training objective using the evidence lower bound (ELBO).
  • The authors show the resulting loss is structurally analogous to offset-noise objectives, with time-dependent coefficients, linking theory to the previously empirical technique.
  • Experiments on controlled synthetic datasets indicate the approach mitigates brightness-related failures and improves performance versus conventional approaches, especially in high-dimensional settings.

Abstract

Diffusion models have become fundamental tools for modeling data distributions in machine learning. Despite their success, these models face challenges when generating data with extreme brightness values, as evidenced by limitations observed in practical large-scale diffusion models. Offset noise has been proposed as an empirical solution to this issue, yet its theoretical basis remains insufficiently explored. In this paper, we propose a novel diffusion model that naturally incorporates additional noise within a rigorous probabilistic framework. Our approach modifies both the forward and reverse diffusion processes, enabling inputs to be diffused into Gaussian distributions with arbitrary mean structures. We derive a loss function based on the evidence lower bound and show that the resulting objective is structurally analogous to that of offset noise, with time-dependent coefficients. Experiments on controlled synthetic datasets demonstrate that the proposed model mitigates brightness-related limitations and achieves improved performance over conventional methods, particularly in high-dimensional settings.