Fine-tuning Factor Augmented Neural Lasso for Heterogeneous Environments

arXiv stat.ML / 4/15/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes FAN-Lasso, a transfer learning framework that fine-tunes pre-learned components for high-dimensional nonparametric regression with variable selection under both covariate and posterior shifts.
  • It introduces a low-rank factor structure to handle dependent high-dimensional covariates and a residual fine-tuning decomposition that represents the target function as a transformation of a frozen source function plus additional terms.
  • The authors derive minimax-optimal excess risk bounds to identify when fine-tuning provides statistical acceleration over single-task learning, based on relative sample sizes and function complexity measures.
  • The framework is positioned as a theoretical lens on parameter-efficient fine-tuning methods, connecting the proposed approach to broader fine-tuning efficiency perspectives.
  • Extensive experiments across multiple shift scenarios show FAN-Lasso outperforming standard baselines and achieving near-oracle performance, including when target-domain sample sizes are severely limited.

Abstract

Fine-tuning is a widely used strategy for adapting pre-trained models to new tasks, yet its methodology and theoretical properties in high-dimensional nonparametric settings with variable selection have not yet been developed. This paper introduces the fine-tuning factor augmented neural Lasso (FAN-Lasso), a transfer learning framework for high-dimensional nonparametric regression with variable selection that simultaneously handles covariate and posterior shifts. We use a low-rank factor structure to manage high-dimensional dependent covariates and propose a novel residual fine-tuning decomposition in which the target function is expressed as a transformation of a frozen source function and other variables to achieve transfer learning and nonparametric variable selection. This augmented feature from the source predictor allows for the transfer of knowledge to the target domain and reduces model complexity there. We derive minimax-optimal excess risk bounds for the fine-tuning FAN-Lasso, characterizing the precise conditions, in terms of relative sample sizes and function complexities, under which fine-tuning yields statistical acceleration over single-task learning. The proposed framework also provides a theoretical perspective on parameter-efficient fine-tuning methods. Extensive numerical experiments across diverse covariate- and posterior-shift scenarios demonstrate that the fine-tuning FAN-Lasso consistently outperforms standard baselines and achieves near-oracle performance even under severe target sample size constraints, empirically validating the derived rates.