Variable Selection Using Relative Importance Rankings
arXiv stat.ML / 4/14/2026
💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper reframes relative importance (RI) analysis—traditionally used for post-hoc model explanation—into a pre-model workflow for feature/variable ranking and filter-based selection.
- It argues RI measures should outperform marginal correlation by capturing both direct and combined predictor effects, thereby accounting for dependencies among variables.
- The authors introduce a new RI metric, CRI.Z, and show it improves computational efficiency versus conventional RI measures.
- Extensive simulations indicate RI-based rankings are more accurate than marginal correlation, particularly under suppressed or weak predictors, and models trained using RI-selected variables are highly competitive versus lasso/relaxed lasso.
- The method also performs well in difficult regimes with clusters of highly correlated predictors and is validated on two high-dimensional gene-expression datasets, with accompanying open-source code.


