Computational bottlenecks for denoising diffusions

arXiv stat.ML / 4/9/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies whether denoising diffusion methods can efficiently sample any target distribution mu for which direct sampling is tractable.
  • It presents evidence that this is not always the case by analyzing a distribution where sampling is easy but the corresponding diffusion drift becomes intractable.
  • Under a conjecture about information-computation gaps in statistical estimation, the authors argue the diffusion drift required for accurate sampling can be computationally prohibitive.
  • They further show there are drifts that are superpolynomially close to the optimum among polynomial-time drifts yet produce generated samples whose distribution remains far from the target.
  • Overall, the work highlights computational bottlenecks and limitations of diffusion-based sampling tied to the complexity of learning/using the drift.

Abstract

Denoising diffusions sample from a probability distribution \mu in \mathbb{R}^d by constructing a stochastic process ({\hat{\boldsymbol x}}_t:t\ge 0) in \mathbb{R}^d such that {\hat{\boldsymbol x}}_0 is easy to sample, but the distribution of \hat{\boldsymbol x}_T at large T approximates \mu. The drift {\boldsymbol m}:\mathbb{R}^d\times\mathbb{R}\to\mathbb{R}^d of this diffusion process is learned my minimizing a score-matching objective. Is every probability distribution \mu, for which sampling is tractable, also amenable to sampling via diffusions? We provide evidence to the contrary by studying a probability distribution \mu for which sampling is easy, but the drift of the diffusion process is intractable -- under a popular conjecture on information-computation gaps in statistical estimation. We show that there exist drifts that are superpolynomially close to the optimum value (among polynomial time drifts) and yet yield samples with distribution that is very far from the target one.