Exact Finite-Sample Variance Decomposition of Subagging: A Spectral Filtering Perspective

arXiv cs.LG / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper provides the first exact finite-sample variance decomposition for subagging using the Hoeffding-ANOVA framework, covering any symmetric base learner without asymptotic or smoothness assumptions.
  • It characterizes subagging as a deterministic low-pass spectral filter that attenuates the c-th order interaction variance by a geometric factor approaching α^c while preserving low-order structural signals.
  • The analysis explains why commonly used default resampling ratios can under-regularize high-capacity interpolating models, which require smaller α to exponentially suppress spurious high-order noise.
  • The authors introduce a complexity-guided adaptive subsampling algorithm that tunes α according to the learner’s complexity spectrum and shows improved generalization versus static baselines in experiments.

Abstract

Standard resampling ratios (e.g., \alpha \approx 0.632) are widely used as default baselines in ensemble learning for three decades. However, how these ratios interact with a base learner's intrinsic functional complexity in finite samples lacks a exact mathematical characterization. We leverage the Hoeffding-ANOVA decomposition to derive the first exact, finite-sample variance decomposition for subagging, applicable to any symmetric base learner without requiring asymptotic limits or smoothness assumptions. We establish that subagging operates as a deterministic low-pass spectral filter: it preserves low-order structural signals while attenuating c-th order interaction variance by a geometric factor approaching \alpha^c. This decoupling reveals why default baselines often under-regularize high-capacity interpolators, which instead require smaller \alpha to exponentially suppress spurious high-order noise. To operationalize these insights, we propose a complexity-guided adaptive subsampling algorithm, empirically demonstrating that dynamically calibrating \alpha to the learner's complexity spectrum consistently improves generalization over static baselines.