Analyzing Shapley Additive Explanations to Understand Anomaly Detection Algorithm Behaviors and Their Complementarity

arXiv stat.ML / 4/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the difficulty of designing truly complementary ensembles for unsupervised anomaly detection, where detectors often share similar cues and generate redundant anomaly scores.
  • It proposes a methodology that uses SHAP (SHapley Additive exPlanations) to characterize each anomaly detector’s decision mechanism by quantifying feature importance attribution patterns.
  • The authors show that detectors with similar SHAP-based explanation profiles tend to output correlated anomaly scores and flag largely overlapping anomalies.
  • In contrast, divergence in explanations is shown to be a reliable indicator of complementary detection behavior, providing a selection criterion different from raw anomaly outputs.
  • The study also finds that explanation diversity alone is not enough; strong individual detector performance is still required, and ensembles built by targeting explanation diversity while preserving quality become more diverse and effective.

Abstract

Unsupervised anomaly detection is a challenging problem due to the diversity of data distributions and the lack of labels. Ensemble methods are often adopted to mitigate these challenges by combining multiple detectors, which can reduce individual biases and increase robustness. Yet building an ensemble that is genuinely complementary remains challenging, since many detectors rely on similar decision cues and end up producing redundant anomaly scores. As a result, the potential of ensemble learning is often limited by the difficulty of identifying models that truly capture different types of irregularities. To address this, we propose a methodology for characterizing anomaly detectors through their decision mechanisms. Using SHapley Additive exPlanations, we quantify how each model attributes importance to input features, and we use these attribution profiles to measure similarity between detectors. We show that detectors with similar explanations tend to produce correlated anomaly scores and identify largely overlapping anomalies. Conversely, explanation divergence reliably indicates complementary detection behavior. Our results demonstrate that explanation-driven metrics offer a different criterion than raw outputs for selecting models in an ensemble. However, we also demonstrate that diversity alone is insufficient; high individual model performance remains a prerequisite for effective ensembles. By explicitly targeting explanation diversity while maintaining model quality, we are able to construct ensembles that are more diverse, more complementary, and ultimately more effective for unsupervised anomaly detection.