BVFLMSP : Bayesian Vertical Federated Learning for Multimodal Survival with Privacy

arXiv stat.ML / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces BVFLMSP, a Bayesian vertical federated learning framework for multimodal time-to-event (survival) prediction that avoids centralized training by keeping modality-specific data at separate clients.
  • BVFLMSP uses a split neural network where each client trains a Bayesian neural network per modality and sends perturbed intermediate representations to a central server for survival risk prediction.
  • Differential privacy is applied to client-side representations to provide formal privacy guarantees against information leakage during federated training.
  • Experiments report improved discrimination versus both single-modality and centralized multimodal baselines (e.g., up to 0.02 higher C-index than MultiSurv) while also producing uncertainty estimates for more reliable decision-making.
  • The study analyzes the performance–privacy tradeoff across different privacy budgets and modality combinations, showing robustness under stricter privacy constraints.

Abstract

Multimodal time-to-event prediction often requires integrating sensitive data distributed across multiple parties, making centralized model training impractical due to privacy constraints. At the same time, most existing multimodal survival models produce single deterministic predictions without indicating how confident the model is in its estimates, which can limit their reliability in real-world decision making. To address these challenges, we propose BVFLMSP, a Bayesian Vertical Federated Learning (VFL) framework for multimodal time-to-event analysis based on a Split Neural Network architecture. In BVFLMSP, each client independently models a specific data modality using a Bayesian neural network, while a central server aggregates intermediate representations to perform survival risk prediction. To enhance privacy, we integrate differential privacy mechanisms by perturbing client side representations before transmission, providing formal privacy guarantees against information leakage during federated training. We first evaluate our Bayesian multimodal survival model against widely used single modality survival baselines and the centralized multimodal baseline MultiSurv. Across multimodal settings, the proposed method shows consistent improvements in discrimination performance, with up to 0.02 higher C-index compared to MultiSurv. We then compare federated and centralized learning under varying privacy budgets across different modality combinations, highlighting the tradeoff between predictive performance and privacy. Experimental results show that BVFLMSP effectively includes multimodal data, improves survival prediction over existing baselines, and remains robust under strict privacy constraints while providing uncertainty estimates.