On the Properties of Feature Attribution for Supervised Contrastive Learning

arXiv cs.AI / 4/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies image classification models trained with supervised contrastive learning (SCL) instead of the usual cross-entropy (CE) objective with an explicit classification layer.
  • It argues that SCL learns a label-informed embedding space that clusters similar samples and separates dissimilar ones, which supports robustness and out-of-distribution detection.
  • The authors empirically evaluate feature attribution quality for models trained with SCL versus contrastive learning (CL), focusing on metrics tied to explanation trustworthiness.
  • They find that SCL-trained networks produce higher-quality feature attribution explanations than CL in terms of faithfulness, complexity, and continuity.
  • The results suggest that choosing training objectives can improve not only accuracy but also transparency and interpretability—important for safety-critical use cases.

Abstract

Most Neural Networks (NNs) for classification are trained using Cross-Entropy as a loss function. This approach requires the model to have an explicit classification layer. However, there exist alternative approaches, such as Contrastive Learning (CL). Instead of explicitly operating a classification, CL has the NN produce an embedding space where projections of similar data are pulled together, while projections of dissimilar data are pushed apart. In the case of Supervised CL (SCL), labels are adopted as similarity criteria, thus creating an embedding space where the projected data points are well-clustered. SCL provides crucial advantages over CE with regard to adversarial robustness and out-of-distribution detection, thus making it a more natural choice in safety-critical scenarios. In the present paper, we empirically show that NNs for image classification trained with SCL present higher-quality feature attribution explanations than CL with regard to faithfulness, complexity, and continuity. These results reinforce previous findings about CL-based approaches when targeting more trustworthy and transparent NNs and can guide practitioners in the selection of training objectives targeting not only accuracy, but also transparency of the models.