Robustness Quantification for Discriminative Models: a New Robustness Metric and its Application to Dynamic Classifier Selection

arXiv cs.LG / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses limitations of existing robustness quantification methods, which typically require generative models and are constrained to certain architectures or discrete feature types.
  • It introduces a new robustness metric designed to work with any probabilistic discriminative classifier and with any kind of input features.
  • The authors show that the proposed metric can effectively separate reliable predictions from unreliable ones, enabling more trustworthy per-instance evaluation.
  • They leverage this separation to develop new strategies for dynamic classifier selection, aiming to choose better-performing models depending on predicted reliability.

Abstract

Among the different possible strategies for evaluating the reliability of individual predictions of classifiers, robustness quantification stands out as a method that evaluates how much uncertainty a classifier could cope with before changing its prediction. However, its applicability is more limited than some of its alternatives, since it requires the use of generative models and restricts the analyses either to specific model architectures or discrete features. In this work, we propose a new robustness metric applicable to any probabilistic discriminative classifier and any type of features. We demonstrate that this new metric is capable of distinguishing between reliable and unreliable predictions, and use this observation to develop new strategies for dynamic classifier selection.