Highly Adaptive Principal Component Regression

arXiv stat.ML / 5/6/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper updates the Highly Adaptive Lasso (HAL) idea by introducing principal-component versions—PCHAL and PCHAR—that reduce the computation burden of HAL/HAR in high-dimensional settings.
  • These new estimators use outcome-blind principal-component reduction applied to the HAL basis, yielding large computational savings while maintaining performance comparable to HAL and HAR in experiments.
  • The authors propose an early-stopped gradient descent variant that acts as a practical form of smooth spectral regularization, avoiding the need to choose a hard cutoff for principal components.
  • They additionally show a theoretical connection: under special conditions, the HAL kernel matches the covariance function of Brownian motion.

Abstract

The Highly Adaptive Lasso (HAL) is a nonparametric regression method that achieves almost dimension-free convergence rates under minimal smoothness assumptions, but its implementation can be computationally prohibitive in high dimensions due to the large design matrix it requires. The Highly Adaptive Ridge (HAR) has been proposed as a related ridge-regularized analogue. Building on both procedures, we introduce the Principal Component Highly Adaptive Lasso (PCHAL) and Principal Component Highly Adaptive Ridge (PCHAR). These estimators use an outcome-blind principal-component reduction of the HAL basis, offering substantial computational gains over HAL while achieving empirical performance comparable to HAL and HAR. We also describe an early-stopped gradient descent variant, which provides a convenient form of smooth spectral regularization without explicitly selecting a hard principal-component cutoff. Finally, we uncover that under special circumstances, the HAL kernel is identical to the covariance function of Brownian motion.