AI Navigate

Seasoning Generative Models for a Generalization Aftertaste

arXiv cs.LG / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper extends a strong-duality result for f-divergences to derive a discriminator-guided refinement recipe that can refine any generative model.
  • The refined models provably exhibit better generalization than their non-refined counterparts, with the generalization gap linked to the Rademacher complexity of the discriminator set used.
  • The framework subsumes a recent score-based diffusion approach (Kim et al., 2022), providing theoretical generalization guarantees for that method.
  • The authors present this work as both a validation of existing methods and a source of new algorithmic avenues, deepening our understanding of generalization in generative modeling.

Abstract

The use of discriminators to train or fine-tune generative models has proven to be a rather successful framework. A notable example is Generative Adversarial Networks (GANs) that minimize a loss incurred by training discriminators along with other paradigms that boost generative models via discriminators that satisfy weak learner constraints. More recently, even diffusion models have shown advantages with some kind of discriminator guidance. In this work, we extend a strong-duality result related to f-divergences which gives rise to a discriminator-guided recipe that allows us to \textit{refine} any generative model. We then show that the refined generative models provably improve generalization, compared to its non-refined counterpart. In particular, our analysis reveals that the gap in generalization is improved based on the Rademacher complexity of the discriminator set used for refinement. Our recipe subsumes a recently introduced score-based diffusion approach (Kim et al., 2022) that has shown great empirical success, however allows us to shed light on the generalization guarantees of this method by virtue of our analysis. Thus, our work provides a theoretical validation for existing work, suggests avenues for new algorithms, and contributes to our understanding of generalization in generative models at large.