Asymptotic Learning Curves for Diffusion Models with Random Features Score and Manifold Data

arXiv cs.LG / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes denoising score matching for diffusion models when the underlying data lies on a low-dimensional manifold and the score is modeled with a random-feature neural network.
  • It provides asymptotically exact high-dimensional expressions for train, test, and score errors, aiming to characterize learning behavior with theory rather than experiments.
  • For linear manifolds, the required sample complexity to learn the score scales with the intrinsic (manifold) dimension instead of the ambient dimension, suggesting a structural efficiency gain.
  • For non-linear manifolds, the advantage from low-dimensional structure weakens, indicating that the benefit depends sensitively on the manifold geometry.
  • Overall, the results suggest diffusion models can exploit structured data, but the type of structure—and how non-linear it is—critically affects learning performance.

Abstract

We study the theoretical behavior of denoising score matching--the learning task associated to diffusion models--when the data distribution is supported on a low-dimensional manifold and the score is parameterized using a random feature neural network. We derive asymptotically exact expressions for the test, train, and score errors in the high-dimensional limit. Our analysis reveals that, for linear manifolds the sample complexity required to learn the score function scales linearly with the intrinsic dimension of the manifold, rather than with the ambient dimension. Perhaps surprisingly, the benefits of low-dimensional structure starts to diminish once we have a non-linear manifold. These results indicate that diffusion models can benefit from structured data; however, the dependence on the specific type of structure is subtle and intricate.