Conformal Margin Risk Minimization: An Envelope Framework for Robust Learning under Label Noise

arXiv cs.LG / 4/9/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Conformal Margin Risk Minimization (CMRM) is proposed as a plug-and-play “envelope” regularization framework that improves any classification loss under label noise without needing privileged information like noise transition matrices or clean subsets.
  • CMRM computes a confidence margin between the observed label and competing labels, then applies a single quantile-calibrated (conformal) threshold per batch to emphasize high-margin samples and suppress likely mislabeled ones.
  • The authors provide a theoretical learning bound for CMRM under arbitrary label noise, relying only on mild regularity assumptions about the margin distribution.
  • Experiments across five base methods and six benchmarks (synthetic and real-world noise) show consistent accuracy gains (up to +3.39%) and smaller conformal prediction sets (up to -20.44%), with no degradation when there is 0% noise.
  • Results suggest CMRM leverages a method-agnostic uncertainty signal—specifically, margin-based conformal calibration—that existing robustness techniques may not fully exploit.

Abstract

Most methods for learning with noisy labels require privileged knowledge such as noise transition matrices, clean subsets or pretrained feature extractors, resources typically unavailable when robustness is most needed. We propose Conformal Margin Risk Minimization (CMRM), a plug-and-play envelope framework that improves any classification loss under label noise by adding a single quantile-calibrated regularization term, with no privileged knowledge or training pipeline modification. CMRM measures the confidence margin between the observed label and competing labels, and thresholds it with a conformal quantile estimated per batch to focus training on high-margin samples while suppressing likely mislabeled ones. We derive a learning bound for CMRM under arbitrary label noise requiring only mild regularity of the margin distribution. Across five base methods and six benchmarks with synthetic and real-world noise, CMRM consistently improves accuracy (up to +3.39%), reduces conformal prediction set size (up to -20.44%) and does not hurt under 0% noise, showing that CMRM captures a method-agnostic uncertainty signal that existing mechanisms did not exploit.