MO-RiskVAE: A Multi-Omics Variational Autoencoder for Survival Risk Modeling in Multiple MyelomaMO-RiskVAE

arXiv cs.LG / 4/9/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces MO-RiskVAE, a multimodal variational autoencoder tailored for survival risk modeling in multiple myeloma by integrating heterogeneous omics with clinical data.
  • Through controlled experiments within an extension of the MyeVAE framework, the authors find that survival-supervised training is most sensitive to the magnitude and structure of latent regularization rather than the particular divergence formulation used.
  • They report that moderately relaxing KL regularization consistently improves survival discrimination, while alternatives like MMD and HSIC show limited benefit unless properly scaled.
  • The study shows that imposing structure on the latent space can improve alignment between learned representations and survival risk gradients, and that a hybrid continuous–discrete latent approach using Gumbel–Softmax can enhance global risk ordering.
  • MO-RiskVAE is presented as a robust model that improves risk stratification over the baseline MyeVAE without adding extra supervision or complex training heuristics.

Abstract

Multimodal variational autoencoders (VAEs) have emerged as a powerful framework for survival risk modeling in multiple myeloma by integrating heterogeneous omics and clinical data. However, when trained under survival supervision, standard latent regularization strategies often fail to preserve prognostically relevant variation, leading to unstable or overly constrained representations. Despite numerous proposed variants, it remains unclear which aspects of latent design fundamentally govern performance in this setting. In this work, we conduct a controlled investigation of latent modeling choices for multimodal survival prediction within a unified extension of the MyeVAE framework. By systematically isolating regularization scale, posterior geometry, and latent space structure under identical architectures and optimization protocols, we show that survival-driven training is primarily sensitive to the magnitude and structure of latent regularization rather than the specific divergence formulation. In particular, moderate relaxation of KL regularization consistently improves survival discrimination, while alternative divergence mechanisms such as MMD and HSIC provide limited benefit without appropriate scaling. We further demonstrate that structuring the latent space can improve alignment between learned representations and survival risk gradients. A hybrid continuous--discrete formulation based on Gumbel--Softmax enhances global risk ordering in the continuous latent subspace, even though stable discrete subtype discovery does not emerge under survival supervision. Guided by these findings, we instantiate a robust multimodal survival model, termed MO-RiskVAE, which consistently improves risk stratification over the original MyeVAE without introducing additional supervision or complex training heuristics.