AI Navigate

A protocol for evaluating robustness to H&E staining variation in computational pathology models

arXiv cs.CV / 3/16/2026

💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The article proposes a three-step protocol to evaluate robustness of computational pathology (CPath) models against variation in H&E staining across laboratories.
  • It constructs a new reference staining library from the PLISM dataset and uses it to assess 306 MSI classification models on the SurGen colorectal cancer dataset, across three feature extractors and six public MSI models under four simulated staining conditions.
  • Classification performance ranged from AUC 0.769 to 0.911 (Δ = 0.142) and robustness ranged from 0.007 to 0.079 (Δ = 0.072), with a weak inverse correlation between robustness and baseline performance (Pearson r = -0.22).
  • The study shows the protocol enables robustness-informed model selection and identifies operational ranges for deployment, with code available at https://github.com/CTPLab/staining-robustness-evaluation.

Abstract

Sensitivity to staining variation remains a major barrier to deploying computational pathology (CPath) models as hematoxylin and eosin (H&E) staining varies across laboratories, requiring systematic assessment of how this variability affects model prediction. In this work, we developed a three-step protocol for evaluating robustness to H&E staining variation in CPath models. Step 1: Select reference staining conditions, Step 2: Characterize test set staining properties, Step 3: Apply CPath model(s) under simulated reference staining conditions. Here, we first created a new reference staining library based on the PLISM dataset. As an exemplary use case, we applied the protocol to assess the robustness properties of 306 microsatellite instability (MSI) classification models on the unseen SurGen colorectal cancer dataset (n=738), including 300 attention-based multiple instance learning models trained on the TCGA-COAD/READ datasets across three feature extractors (UNI2-h, H-Optimus-1, Virchow2), alongside six public MSI classification models. Classification performance was measured as AUC, and robustness as the min-max AUC range across four simulated staining conditions (low/high H&E intensity, low/high H&E color similarity). Across models and staining conditions, classification performance ranged from AUC 0.769-0.911 (\Delta = 0.142). Robustness ranged from 0.007-0.079 (\Delta = 0.072), and showed a weak inverse correlation with classification performance (Pearson r=-0.22, 95% CI [-0.34, -0.11]). Thus, we show that the proposed evaluation protocol enables robustness-informed CPath model selection and provides insight into performance shifts across H&E staining conditions, supporting the identification of operational ranges for reliable model deployment. Code is available at https://github.com/CTPLab/staining-robustness-evaluation .