PAC-Bayes Bounds for Gibbs Posteriors via Singular Learning Theory

arXiv stat.ML / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper derives explicit non-asymptotic PAC-Bayes generalization bounds for Gibbs posteriors, which are data-dependent parameter distributions formed by exponentially tilting a prior using empirical risk.
  • It replaces classical worst-case, metric-entropy-based complexity controls with posterior-averaged risk bounds, making the approach applicable to overparameterized models and able to adapt to data structure and intrinsic complexity.
  • The authors analyze a marginal-type integral in the bound using singular learning theory, enabling explicit and practically interpretable characterizations of posterior risk.
  • Experiments/applications to low-rank matrix completion and ReLU neural network regression and classification produce bounds that are analytically tractable and substantially tighter than classical complexity-based PAC-Bayes/learning-theory bounds.
  • Overall, the work demonstrates that PAC-Bayes analysis can provide precise finite-sample generalization guarantees for modern overparameterized and singular learning settings.

Abstract

We derive explicit non-asymptotic PAC-Bayes generalization bounds for Gibbs posteriors, that is, data-dependent distributions over model parameters obtained by exponentially tilting a prior with the empirical risk. Unlike classical worst-case complexity bounds based on uniform laws of large numbers, which require explicit control of the model space in terms of metric entropy (integrals), our analysis yields posterior-averaged risk bounds that can be applied to overparameterized models and adapt to the data structure and the intrinsic model complexity. The bound involves a marginal-type integral over the parameter space, which we analyze using tools from singular learning theory to obtain explicit and practically meaningful characterizations of the posterior risk. Applications to low-rank matrix completion and ReLU neural network regression and classification show that the resulting bounds are analytically tractable and substantially tighter than classical complexity-based bounds. Our results highlight the potential of PAC-Bayes analysis for precise finite-sample generalization guarantees in modern overparameterized and singular models.