A comparative analysis of machine learning models in SHAP analysis

arXiv cs.LG / 4/9/2026

💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper argues that black-box machine learning models are increasingly common but are often hard to interpret, motivating the use of SHAP (SHapley Additive exPlanations) for feature-level explanation of predictions.
  • It notes that SHAP value interpretation depends on the specific underlying model, which means there is no single universal SHAP analysis procedure.
  • The authors provide a comparative investigation of SHAP analysis across different machine learning models and datasets to characterize the nuances in how SHAP outputs should be interpreted.
  • The work includes a new generalization of the waterfall plot for multi-class classification problems to better visualize per-class/per-sample contribution breakdowns.

Abstract

In this growing age of data and technology, large black-box models are becoming the norm due to their ability to handle vast amounts of data and learn incredibly complex data patterns. The deficiency of these methods, however, is their inability to explain the prediction process, making them untrustworthy and their use precarious in high-stakes situations. SHapley Additive exPlanations (SHAP) analysis is an explainable AI method growing in popularity for its ability to explain model predictions in terms of the original features. For each sample and feature in the data set, an associated SHAP value quantifies the contribution of that feature to the prediction of that sample. Analysis of these SHAP values provides valuable insight into the model's decision-making process, which can be leveraged to create data-driven solutions. The interpretation of these SHAP values, however, is model-dependent, so there does not exist a universal analysis procedure. To aid in these efforts, we present a detailed investigation of SHAP analysis across various machine learning models and data sets. In uncovering the details and nuance behind SHAP analysis, we hope to empower analysts in this less-explored territory. We also present a novel generalization of the waterfall plot to the multi-classification problem.