Stability of a Generalized Debiased Lasso with Applications to Resampling-Based Variable Selection

arXiv stat.ML / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a generalized debiased Lasso estimator built on a stability principle, providing an update formula when one column of the design matrix is perturbed.
  • Under sub-Gaussian design assumptions and well-conditioned covariance, the approximation is shown to be asymptotically accurate for all but a vanishing fraction of coordinates in the proportional growth regime.
  • The theoretical results use concentration and anti-concentration techniques to control remainder terms and prevent sign changes from dominating the error.
  • The authors note that deriving stronger distributional limits (such as Gaussianity) under similar assumptions is still an open problem.
  • As an application, the update-based approximation can substantially reduce the computational cost of resampling-based variable selection methods, including conditional randomization tests and a local knockoff filter.

Abstract

We propose a generalized debiased Lasso estimator based on a stability principle. When a single column of the design matrix is perturbed, the estimator admits a simple update formula that can be computed from the original solution. Under sub-Gaussian designs with well-conditioned covariance, this approximation is asymptotically accurate for all but a vanishing fraction of coordinates in the proportional growth regime. The proof relies on concentration and anti-concentration arguments to control error terms and sign changes. In contrast, establishing comparable distributional limits (e.g., Gaussianity) under similar assumptions remains open. As an application, we show that the approximation significantly reduces the computational cost of resampling-based variable selection procedures, including the conditional randomization test and a local knockoff filter.