Tucker Diffusion Model for High-dimensional Tensor Generation

arXiv stat.ML / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a Tucker diffusion model to generate structured high-dimensional tensor-valued data with a target distribution, extending diffusion models beyond multi-linear tensor observations.
  • It shows that, under a low Tucker-rank assumption, the diffusion model’s score function can be decomposed in a structured way and estimated efficiently using a specialized Tucker-Unet architecture.
  • The authors provide a theoretical result that the generated tensor distribution converges to the true distribution at a rate tied to the maximum tensor mode dimension, improving over naive vectorized methods with a product-dimension dependence.
  • Experiments on synthetic and real-world tensor generation tasks indicate the proposed approach can match or outperform existing methods while reducing training and sampling costs.

Abstract

Statistical inference on large-dimensional tensor data has been extensively studied in the literature and widely used in economics, biology, machine learning, and other fields, but how to generate a structured tensor with a target distribution is still a new problem. As profound AI generators, diffusion models have achieved remarkable success in learning complex distributions. However, their extension to generating multi-linear tensor-valued observations remains underexplored. In this work, we propose a novel Tucker diffusion model for learning high-dimensional tensor distributions. We show that the score function admits a structured decomposition under the low Tucker rank assumption, allowing it to be both accurately approximated and efficiently estimated using a carefully tailored tensor-shaped architecture named Tucker-Unet. Furthermore, the distribution of generated tensors, induced by the estimated score function, converges to the true data distribution at a rate depending on the maximum of tensor mode dimensions, thereby offering a clear theoretical advantage over the naive vectorized approach, which has a product dependence. Empirically, compared to existing approaches, the Tucker diffusion model demonstrates strong practical potential in synthetic and real-world tensor generation tasks, achieving comparable and sometimes even superior statistical performance with significantly reduced training and sampling costs.