Measuring the Sensitivity of Classification Models with the Error Sensitivity Profile

arXiv cs.LG / 4/29/2026

📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper introduces the Error Sensitivity Profile (ESP) to measure how sensitive classification model performance is to errors in individual features or combinations of features.
  • ESP is designed to help prioritize data-cleaning work by focusing on the error types and specific features most likely to harm model performance.
  • To compute ESP in practice, the authors release an integrated tool suite called \\dirty.
  • Experiments across two common datasets and 14 classification models show that performance drops are not always explained by simple correlations with the target variable.
  • Overall, the work provides a data-centric way to diagnose which data errors matter most for downstream classification accuracy.

Abstract

The quality of training data is critical to the performance of machine learning models. In this paper, the Error Sensitivity Profile (ESP) is proposed. It quantifies the sensitivity of model performance to errors in a single feature or in multiple features. By leveraging ESP, data-cleaning efforts can be prioritized based on error types and features most likely to affect model performance. To support the computation of this metric, an integrated suite of tools, called \dirty, is created. We conduct an extensive experimental study on two widely used datasets using 14 classification models, revealing that performance degradation is not always predictable from simple correlations with the target variable.