Regional Explanations: Bridging Local and Global Variable Importance

arXiv stat.ML / 4/14/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper evaluates two common local explanation methods, Local Shapley Values and LIME, and argues they can fail to reliably identify features that are truly locally important for a particular prediction.
  • It claims a principled local attribution method should avoid assigning importance to features that do not affect the model output or are statistically independent of features that drive the prediction.
  • The authors show that both Local SV and LIME can violate this principle even in idealized settings with exact computation and independent features.
  • To fix this, they propose R-LOCO, which partitions the input space into regions with similar feature-importance behavior and then uses global attribution methods within those regions to produce more faithful local attributions.
  • The approach is designed to reduce local explanation instability while retaining instance-specific details that global attribution methods often miss.

Abstract

We analyze two widely used local attribution methods, Local Shapley Values and LIME, which aim to quantify the contribution of a feature value x_i to a specific prediction f(x_1, \dots, x_p). Despite their widespread use, we identify fundamental limitations in their ability to reliably detect locally important features, even under ideal conditions with exact computations and independent features. We argue that a sound local attribution method should not assign importance to features that neither influence the model output (e.g., features with zero coefficients in a linear model) nor exhibit statistical dependence with functionality-relevant features. We demonstrate that both Local SV and LIME violate this fundamental principle. To address this, we propose R-LOCO (Regional Leave Out COvariates), which bridges the gap between local and global explanations and provides more accurate attributions. R-LOCO segments the input space into regions with similar feature importance characteristics. It then applies global attribution methods within these regions, deriving an instance's feature contributions from its regional membership. This approach delivers more faithful local attributions while avoiding local explanation instability and preserving instance-specific detail often lost in global methods.