AI Navigate

Fair Lung Disease Diagnosis from Chest CT via Gender-Adversarial Attention Multiple Instance Learning

arXiv cs.CV / 3/16/2026

📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • A fairness-aware MIL framework for multi-class lung disease diagnosis from chest CT volumes is proposed for a CVPR 2026 workshop, focusing on four classes (Healthy, COVID-19, Adenocarcinoma, Squamous Cell Carcinoma) and evaluating by average per-gender macro F1 to punish gender inequity.
  • The model uses Attention-based MIL on a ConvNeXt backbone to identify diagnostically relevant slices without slice-level supervision, with a Gradient Reversal Layer to adversarially suppress gender-predictive information in the representation.
  • The training setup includes focal loss with label smoothing, stratified cross-validation over joint (class, gender) strata, and targeted oversampling of underrepresented subgroups to address demographic imbalance.
  • Inference uses five-fold ensemble of checkpoints with horizontal-flip test-time augmentation and soft logit voting, plus out-of-the-fold threshold optimization, achieving mean validation score 0.685 (std 0.030) and best fold 0.759.
  • The training and inference code is publicly available at https://github.com/ADE-17/cvpr-fair-chest-ct.

Abstract

We present a fairness-aware framework for multi-class lung disease diagnosis from chest CT volumes, developed for the Fair Disease Diagnosis Challenge at the PHAROS-AIF-MIH Workshop (CVPR 2026). The challenge requires classifying CT scans into four categories -- Healthy, COVID-19, Adenocarcinoma, and Squamous Cell Carcinoma -- with performance measured as the average of per-gender macro F1 scores, explicitly penalizing gender-inequitable predictions. Our approach addresses two core difficulties: the sparse pathological signal across hundreds of slices, and a severe demographic imbalance compounded across disease class and gender. We propose an attention-based Multiple Instance Learning (MIL) model on a ConvNeXt backbone that learns to identify diagnostically relevant slices without slice-level supervision, augmented with a Gradient Reversal Layer (GRL) that adversarially suppresses gender-predictive structure in the learned scan representation. Training incorporates focal loss with label smoothing, stratified cross-validation over joint (class, gender) strata, and targeted oversampling of the most underrepresented subgroup. At inference, all five-fold checkpoints are ensembled with horizontal-flip test-time augmentation via soft logit voting and out-of-the-fold threshold optimization for robustness. Our model achieves a mean validation competition score of 0.685 (std - 0.030), with the best single fold reaching 0.759. All training and inference code is publicly available at https://github.com/ADE-17/cvpr-fair-chest-ct