LOCO Feature Importance Inference without Data Splitting via Minipatch Ensembles
arXiv stat.ML / 3/24/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a mostly model-agnostic, distribution-free framework for feature importance inference using LOCO-style occlusion/leave-one-covariate-out ideas, but without requiring data splitting.
- The core method uses “minipatch ensembles,” combining random observation and feature subsampling so inference can be performed with trained ensembles, avoiding model refitting and held-out test data.
- It argues for both computational and statistical efficiency while addressing interpretability issues commonly caused by data-splitting procedures.
- The authors provide theoretical results showing asymptotic validity of confidence intervals under mild assumptions, even when training and inference reuse the same data.
- Practical theory-backed remedies are included for challenges like vanishing variance for null features and performing inference after data-driven hyperparameter tuning.
Related Articles

Interactive Web Visualization of GPT-2
Reddit r/artificial
Stop Treating AI Interview Fraud Like a Proctoring Problem
Dev.to
[R] Causal self-attention as a probabilistic model over embeddings
Reddit r/MachineLearning
The 5 software development trends that actually matter in 2026 (and what they mean for your startup)
Dev.to
InVideo AI Review: Fast Finished
Dev.to