Federated Distillation for Whole Slide Image via Gaussian-Mixture Feature Alignment and Curriculum Integration

arXiv cs.CV / 5/4/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces FedHD, a federated learning framework for whole slide image (WSI) digital pathology that tackles cross-institution heterogeneity in MIL architectures and feature extractors.
  • FedHD replaces parameter sharing by having each client distill semantically rich synthetic features using local Gaussian-mixture feature alignment tied to the real WSI feature distribution.
  • To avoid loss of diagnostic diversity, it uses a one-to-one distillation scheme that creates a synthetic counterpart for each real slide, preventing over-compression.
  • The method adds curriculum-based integration that gradually incorporates cross-site synthetic features into local training when performance plateaus, improving stable collaboration.
  • An optional interpretation module reconstructs pseudo-patches from synthetic embeddings for transparency, and experiments on TCGA-IDH and CAMELYON16/17 show consistent gains over prior federated and distillation baselines.

Abstract

Federated learning (FL) offers a promising framework for collaborative digital pathology by enabling model training across institutions. However, real-world deployments face heterogeneity arising from diverse multiple instance learning (MIL) architectures and heterogeneous feature extractors across institutions. We propose FedHD, a novel FL framework that performs local Gaussian-mixture feature alignment tailored for WSI analysis. Instead of exchanging model parameters, each client independently distills semantically rich synthetic feature representations aligned with the distribution of real WSIs. To preserve diagnostic diversity, FedHD adopts a one-to-one distillation strategy, generating a synthetic counterpart for each real slide to avoid over-compression. During federation, a curriculum-based integration strategy progressively incorporates cross-site synthetic features into local training once performance plateaus. Furthermore, an optional interpretation module reconstructs pseudo-patches from synthetic embeddings, enhancing transparency. FedHD is architecture-agnostic, privacy-preserving, and supports personalized yet collaborative training across diverse institutions. Experiments on TCGA-IDH, CAMELYON16, and CAMELYON17 show that FedHD consistently outperforms state-of-the-art federated and distillation baselines.