AICO: Feature Significance Tests for Supervised Learning

arXiv stat.ML / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces AICO, a framework for statistically testing feature importance by masking individual features and measuring how predictive performance changes.
  • AICO is designed to provide exact, finite-sample feature p-values and confidence intervals using a non-asymptotic hypothesis testing procedure.
  • Unlike many existing interpretability approaches, AICO does not require retraining, surrogate modeling, or distributional assumptions, aiming to stay practical for large modern models.
  • The authors report that AICO works effectively in both controlled experiments and real applications (e.g., credit scoring and mortgage-behavior prediction), reliably identifying the features driving model behavior.
  • The method is positioned as a way to improve transparency, fairness/accountability checks, and policy confidence in model-based decisions by grounding interpretability in statistical guarantees.

Abstract

Machine learning is central to modern science, industry, and policy, yet its predictive power often comes at the cost of transparency: we rarely know which input features truly drive a model's predictions. Without such understanding, researchers cannot draw reliable conclusions, practitioners cannot ensure fairness or accountability, and policymakers cannot trust or govern model-based decisions. Existing tools for assessing feature influence are limited; most lack statistical guarantees, and many require costly retraining or surrogate modeling, making them impractical for large modern models. We introduce AICO, a broadly applicable framework that turns model interpretability into an efficient statistical exercise. AICO tests whether each feature genuinely improves predictive performance by masking its information and measuring the resulting change. The method provides exact, finite-sample feature p-values and confidence intervals for feature importance through a simple, non-asymptotic hypothesis testing procedure. It requires no retraining, surrogate modeling, or distributional assumptions, making it feasible for large-scale algorithms. In both controlled experiments and real applications, from credit scoring to mortgage-behavior prediction, AICO reliably identifies the variables that drive model behavior, providing a scalable and statistically principled path toward transparent and trustworthy machine learning.