GraphPL: Leveraging GNN for Efficient and Robust Modalities Imputation in Patchwork Learning

arXiv cs.LG / 4/29/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces patchwork learning for distributed multi-modal settings where different clients have access to different subsets of modalities.
  • It targets unsupervised imputation of missing modalities per client, rather than assuming complete modality information is always available.
  • Existing approaches are argued to underutilize available modality information by depending on only a subset of observed modalities.
  • GraphPL is proposed as a graph-neural-network-based method that flexibly integrates all observed modalities and improves robustness to noisy inputs.
  • Experiments report state-of-the-art results on benchmarks and strong downstream performance on real-world distributed electronic health records, including disease prediction driven by better modality imputation.

Abstract

Current research on distributed multi-modal learning typically assumes that clients can access complete information across all modalities, which may not hold in practice. In this paper, we explore patchwork learning, in which the modalities available to different clients vary, and the objective is to impute the missing modalities for each client in an unsupervised manner. Existing methods are shown not to fully utilize the modality information as they tend to rely on only a subset of the observed modalities. To address this issue, we propose GraphPL, which combines graph neural networks with patchwork learning to flexibly integrate all observed modalities and remains robust with noisy inputs. Experimental results show that GraphPL achieves SOTA performance on benchmark datasets. Our results on real-world distributed electronic health record dataset show GraphPL learns strong downstream features and enables tasks like disease prediction via superior modality imputation.