Towards Rigorous Explainability by Feature Attribution

arXiv cs.AI / 4/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that traditional non-symbolic explanation methods for machine learning models often lack rigor and can mislead decision-makers, especially in high-stakes settings.
  • It highlights a concrete example of the rigor gap: the use of Shapley values in XAI, commonly implemented via tools like SHAP.
  • The work surveys ongoing efforts to replace or complement non-rigorous approaches with more rigorous symbolic explainability methods.
  • The focus is on producing more dependable assignments of relative feature importance by using symbolic XAI techniques rather than relying solely on popular non-symbolic attribution methods.

Abstract

For around a decade, non-symbolic methods have been the option of choice when explaining complex machine learning (ML) models. Unfortunately, such methods lack rigor and can mislead human decision-makers. In high-stakes uses of ML, the lack of rigor is especially problematic. One prime example of provable lack of rigor is the adoption of Shapley values in explainable artificial intelligence (XAI), with the tool SHAP being a ubiquitous example. This paper overviews the ongoing efforts towards using rigorous symbolic methods of XAI as an alternative to non-rigorous non-symbolic approaches, concretely for assigning relative feature importance.