Instance-Level Costs for Nuanced Classifier Evaluation

arXiv cs.LG / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a new evaluation metric, normalized excess cost (NEC), to weight classification mistakes by per-example costs instead of treating all errors equally.
  • NEC can be derived from sources such as annotator vote margins, distance to decision thresholds, or confidence ratings, and it reduces to standard error rate when costs are uniform.
  • Experiments across text, image, and tabular benchmarks show NEC is often much lower than error rate, indicating that many errors occur on ambiguous and relatively low-cost examples.
  • Cost-sensitive training methods (e.g., loss weighting, sampling, or cost regression) produce inconsistent results, with clear gains mainly when costs are predictable from input features, as demonstrated in a synthetic control.
  • The authors present a practical framework for deriving and evaluating instance-level misclassification costs, even in settings where cost-sensitive training provides limited improvement.

Abstract

Standard classification treats all errors equally, but in content moderation, medical screening, and safety-critical applications, mistakes on clear-cut cases are far more costly than errors on ambiguous ones. We propose normalized excess cost (NEC), a metric that weights classification errors by per-example costs and reduces to standard error rate when costs are uniform. Costs can derive from annotator vote margins, distance from decision thresholds, or confidence ratings. Across text, image, and tabular benchmarks, we find that NEC is often substantially lower than error rate -- models with 5\% error rate can achieve 1.8\% NEC -- revealing that most mistakes concentrate on ambiguous, low-cost examples. However, incorporating costs into training via loss weighting, sampling strategies, or regression yields inconsistent benefits: improvements appear only when costs are predictable from input features, as in our synthetic control, while real-world datasets show mixed or negligible gains. Our framework provides a practical methodology for deriving and evaluating instance-level misclassification costs, even when cost-sensitive training offers limited benefit.