Conditional Imputation for Within-Modality Missingness in Multi-Modal Federated Learning

arXiv cs.LG / 4/28/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper addresses a common challenge in multimodal federated learning for clinical data: within-modality missingness caused by sensor dropouts or irregular sampling.
  • It criticizes existing approaches that rely on architectural alignment or missing embeddings because they may not recover the true underlying data distribution, hurting performance.
  • The proposed framework, CondI, uses conditional diffusion models in a two-phase training process: first imputing missing temporal components with multimodal context, then training modality-specific extractors and joint embedding spaces.
  • Inference runs the imputed raw data through the learned extractors to produce more robust features for downstream tasks, improving resilience to severe incompleteness.
  • Experiments on PTB-XL, SLEEP-EDF, and MIMIC-IV show CondI achieves results comparable to state-of-the-art baselines, and the authors provide code on GitHub.

Abstract

Multimodal Federated Learning (MMFL) enables privacy-preserving collaborative training, but real-world clinical applications often suffer from within-modality missingness caused by sensor intermittency or irregular sampling. Existing methods implicitly represent unobserved data via architectural alignment or missing embeddings, often failing to recover the true distribution and yielding sub-optimal performance. We propose CondI, a federated framework explicitly addressing this missingness using conditional diffusion models. CondI employs a two-phase training pipeline: first, imputing unobserved temporal components using available multimodal context and conditional embeddings; second, optimizing modality-specific extractors and joint embedding spaces. During inference, imputed raw data pass through trained extractors to generate robust features, providing a holistic representation for downstream tasks. Explicit data imputation ensures models operate on complete semantic structures, significantly enhancing resilience against severe data incompleteness. Experiments on three clinical datasets (PTB-XL, SLEEP-EDF, MIMIC-IV) demonstrate CondI achieves comparable results to state-of-the-art baselines. Code: https://github.com/ZhengWugeng/CondI