Attention-based multiple instance learning for predominant growth pattern prediction in lung adenocarcinoma wsi using foundation models

arXiv cs.CV / 4/24/2026

📰 NewsModels & Research

Key Points

  • The study tackles lung adenocarcinoma grading by predicting the predominant growth pattern directly at the whole-slide level, aiming to improve prognosis-relevant classification.
  • It introduces an attention-based multiple instance learning (ABMIL) framework that aggregates patch-level features using attention, reducing the need for exhaustive patch-level labels.
  • The method uses pretrained pathology foundation models as patch encoders, which can be either frozen or fine-tuned on annotated patches to extract discriminative representations.
  • Experimental results indicate that fine-tuning improves accuracy, with Prov-GigaPath combined with ABMIL achieving the top agreement (kappa = 0.699).
  • The authors plan to extend the approach to predict the full distribution of growth patterns and to validate the model on external cohorts.

Abstract

Lung adenocarcinoma (LUAD) grading depends on accurately identifying growth patterns, which are indicators of prognosis and can influence treatment decisions. Common deep learning approaches to determine the predominant pattern rely on patch-level classification or segmentation, requiring extensive annotations. This study proposes an attention-based multiple instance learning (ABMIL) framework to predict the predominant LUAD growth pattern at the whole slide level to reduce annotation burden. Our approach integrates pretrained pathology foundation models as patch encoders, used either frozen or fine-tuned on annotated patches, to extract discriminative features that are aggregated through attention mechanisms. Experiments show that fine-tuned encoders improve performance, with Prov-GigaPath achieving the highest agreement (\k{appa} = 0.699) under ABMIL. Compared to simple patch-aggregation baselines, ABMIL yields more robust predictions by leveraging slide-level supervision and spatial attention. Future work will extend this framework to estimate the full distribution of growth patterns and validate performance on external cohorts.