LOCO Feature Importance Inference without Data Splitting via Minipatch Ensembles
arXiv stat.ML / 2026/3/24
💬 オピニオンIdeas & Deep AnalysisModels & Research
要点
- The paper proposes a mostly model-agnostic, distribution-free framework for feature importance inference using LOCO-style occlusion/leave-one-covariate-out ideas, but without requiring data splitting.
- The core method uses “minipatch ensembles,” combining random observation and feature subsampling so inference can be performed with trained ensembles, avoiding model refitting and held-out test data.
- It argues for both computational and statistical efficiency while addressing interpretability issues commonly caused by data-splitting procedures.
- The authors provide theoretical results showing asymptotic validity of confidence intervals under mild assumptions, even when training and inference reuse the same data.
- Practical theory-backed remedies are included for challenges like vanishing variance for null features and performing inference after data-driven hyperparameter tuning.




