Spectral-Transport Stability and Benign Overfitting in Interpolating Learning

arXiv stat.ML / 4/13/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a theoretical generalization framework for interpolating (zero training error) learning, aiming to explain how overparameterized models can still generalize well.
  • It introduces a “spectral-transport stability” approach that bounds excess risk using the data distribution’s spectral geometry, sensitivity of the learning rule to single-sample changes, and structure/alignment of label noise.
  • The authors define a scale-dependent “Fredriksson index” that unifies effective dimension, transport stability, and noise alignment into a single complexity parameter for interpolating estimators.
  • Finite-sample risk bounds are proved, and a sharp benign-overfitting criterion is characterized by the vanishing of the index along admissible spectral scales.
  • For polynomial spectral decay and a specialized case of polynomial-spectrum linear interpolation, the paper derives explicit phase-transition rates and shows how optimization dynamics can implicitly select interpolating solutions with minimal spectral-transport energy.

Abstract

We develop a theoretical framework for generalization in the interpolating regime of statistical learning. The central question is why highly overparameterized estimators can attain zero empirical risk while still achieving nontrivial predictive accuracy, and how to characterize the boundary between benign and destructive overfitting. We introduce a spectral-transport stability framework in which excess risk is controlled jointly by the spectral geometry of the data distribution, the sensitivity of the learning rule under single-sample replacement, and the alignment structure of label noise. This leads to a scale-dependent Fredriksson index that combines effective dimension, transport stability, and noise alignment into a single complexity parameter for interpolating estimators. We prove finite-sample risk bounds, establish a sharp benign-overfitting criterion through the vanishing of the index along admissible spectral scales, and derive explicit phase-transition rates under polynomial spectral decay. For a model-specific specialization, we obtain an explicit theorem for polynomial-spectrum linear interpolation, together with a proof of the resulting rate. The framework also clarifies implicit regularization by showing how optimization dynamics can select interpolating solutions of minimal spectral-transport energy. These results connect algorithmic stability, double descent, benign overfitting, operator-theoretic learning theory, and implicit bias within a unified structural account of modern interpolation.