AI Navigate

Longitudinal Risk Prediction in Mammography with Privileged History Distillation

arXiv cs.LG / 3/18/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Privileged History Distillation (PHD), a method that uses full longitudinal mammography history as privileged information during training and distills its prognostic value into a student model that only requires the current exam at inference.
  • It employs horizon-specific teachers trained on complete histories to specialize in particular prediction horizons, enabling horizon-aware risk assessment when prior exams are unavailable at deployment.
  • The approach is validated on the CSAW-CC dataset with multi-year cancer outcomes, showing improved time-dependent AUC over no-history baselines and comparable performance to full-history models.
  • By addressing missing or irregular longitudinal history in real-world clinics, PHD enhances the practicality and robustness of longitudinal mammography risk prediction for deployment.

Abstract

Breast cancer remains a leading cause of cancer-related mortality worldwide. Longitudinal mammography risk prediction models improve multi-year breast cancer risk prediction based on prior screening exams. However, in real-world clinical practice, longitudinal histories are often incomplete, irregular, or unavailable due to missed screenings, first-time examinations, heterogeneous acquisition schedules, or archival constraints. The absence of prior exams degrades the performance of longitudinal risk models and limits their practical applicability. While substantial longitudinal history is available during training, prior exams are commonly absent at test time. In this paper, we address missing history at inference time and propose a longitudinal risk prediction method that uses mammography history as privileged information during training and distills its prognostic value into a student model that only requires the current exam at inference time. The key idea is a privileged multi-teacher distillation scheme with horizon-specific teachers: each teacher is trained on the full longitudinal history to specialize in one prediction horizon, while the student receives only a reconstructed history derived from the current exam. This allows the student to inherit horizon-dependent longitudinal risk cues without requiring prior screening exams at deployment. Our new Privileged History Distillation (PHD) method is validated on a large longitudinal mammography dataset with multi-year cancer outcomes, CSAW-CC, comparing full-history and no-history baselines to their distilled counterparts. Using time-dependent AUC across horizons, our privileged history distillation method markedly improves the performance of long-horizon prediction over no-history models and is comparable to that of full-history models, while using only the current exam at inference time.