Making Conformal Predictors Robust in Healthcare Settings: a Case Study on EEG Classification

arXiv stat.ML / 5/1/2026

💬 OpinionTools & Practical UsageModels & Research

Key Points

  • The paper addresses the need to quantify uncertainty in clinical diagnosis models and highlights conformal prediction as a method with theoretical coverage guarantees.
  • It shows that standard conformal predictors can fail in healthcare because patient distribution shifts break the i.i.d. assumptions, resulting in poor coverage.
  • Using EEG seizure classification as a case study with known distribution-shift and label uncertainty, the authors evaluate multiple conformal prediction approaches.
  • The study finds that personalized calibration strategies can improve coverage by more than 20 percentage points while keeping prediction set sizes comparable.
  • The work provides an open-source implementation via PyHealth to support adoption in healthcare AI workflows.

Abstract

Quantifying uncertainty in clinical predictions is critical for high-stakes diagnosis tasks. Conformal prediction offers a principled approach by providing prediction sets with theoretical coverage guarantees. However, in practice, patient distribution shifts violate the i.i.d. assumptions underlying standard conformal methods, leading to poor coverage in healthcare settings. In this work, we evaluate several conformal prediction approaches on EEG seizure classification, a task with known distribution shift challenges and label uncertainty. We demonstrate that personalized calibration strategies can improve coverage by over 20 percentage points while maintaining comparable prediction set sizes. Our implementation is available via PyHealth, an open-source healthcare AI framework: https://github.com/sunlabuiuc/PyHealth.