The Information Dynamics of Generative Diffusion

arXiv stat.ML / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes an integrated framework for generative diffusion models by linking information-theoretic, dynamical, and thermodynamic viewpoints into a single explanation of how generation proceeds.
  • It shows that the rate of conditional entropy production ("generative bandwidth") during generation is determined by the expected divergence of the score function’s vector field.
  • The work connects this divergence to trajectory branching and “generative bifurcations,” interpreting them as symmetry-breaking phase transitions in the model’s energy landscape.
  • Beyond averages across ensembles, it demonstrates that symmetry-breaking outcomes can be identified via peaks in the variance of pathwise conditional entropy, reflecting trajectory-level heterogeneity in resolving uncertainty.
  • Overall, the authors characterize generative diffusion as controlled, noise-induced symmetry breaking, where the score function behaves like a dynamic nonlinear filter regulating both information-flow rate and variability from noise to data.

Abstract

Generative diffusion models have emerged as a powerful class of models in machine learning, yet a unified theoretical understanding of their operation is still developing. This paper provides an integrated perspective on generative diffusion by connecting the information-theoretic, dynamical, and thermodynamic aspects. We demonstrate that the rate of conditional entropy production during generation (i.e., the generative bandwidth) is directly governed by the expected divergence of the score function's vector field. This divergence, in turn, is linked to the branching of trajectories and generative bifurcations, which we characterize as symmetry-breaking phase transitions in the energy landscape. Beyond ensemble averages, we demonstrate that symmetry-breaking decisions are revealed by peaks in the variance of pathwise conditional entropy, capturing heterogeneity in how individual trajectories resolve uncertainty. Together, these results establish generative diffusion as a process of controlled, noise-induced symmetry breaking, in which the score function acts as a dynamic nonlinear filter that regulates both the rate and variability of information flow from noise to data.