Optimizing Diffusion Priors with a Single Observation

arXiv cs.LG / 4/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Diffusion priors can produce strong posterior samples, but existing models often reflect biases and errors from limited or simulated training data.
  • Current fine-tuning methods typically require many observations across different forward operators, which is hard to obtain and can cause overfitting when data are scarce.
  • The paper proposes tuning a diffusion prior using only a single observation by forming a product-of-experts prior from existing diffusion priors and choosing exponents that maximize Bayesian evidence.
  • Experiments on real-world inverse problems, including black hole imaging and text-conditioned image deblurring, show the best evidence can come from combinations that go beyond priors trained on a single dataset.
  • The exponent-weighted combination allows posterior sampling from both tempered and combined diffusion models, producing more flexible priors that improve the reliability of posterior image distributions.

Abstract

While diffusion priors generate high-quality posterior samples across many inverse problems, they are often trained on limited training sets or purely simulated data, thus inheriting the errors and biases of these underlying sources. Current approaches to finetuning diffusion models rely on a large number of observations with varying forward operators, which can be difficult to collect for many applications, and thus lead to overfitting when the measurement set is small. We propose a method for tuning a prior from only a single observation by combining existing diffusion priors into a single product-of-experts prior and identifying the exponents that maximize the Bayesian evidence. We validate our method on real-world inverse problems, including black hole imaging, where the true prior is unknown a priori, and image deblurring with text-conditioned priors. We find that the evidence is often maximized by priors that extend beyond those trained on a single dataset. By generalizing the prior through exponent weighting, our approach enables posterior sampling from both tempered and combined diffusion models, yielding more flexible priors that improve the trustworthiness of the resulting posterior image distribution.