DMMRL: Disentangled Multi-Modal Representation Learning via Variational Autoencoders for Molecular Property Prediction
arXiv cs.LG / 2026/3/24
📰 ニュースSignals & Early TrendsIdeas & Deep AnalysisModels & Research
要点
- The paper introduces DMMRL, a variational autoencoder-based method for disentangling molecular representations into shared (structure-relevant) and private (modality-specific) latent spaces to address entangled structure-property factors.
- It improves cross-modal learning by using orthogonality and alignment regularizations to encourage statistical independence and consistency across graphs, sequences, and geometries rather than naive concatenation.
- A gated attention fusion module adaptively combines shared representations, aiming to capture richer inter-modal dependencies for molecular property prediction.
- Experiments on seven benchmark datasets show DMMRL outperforming existing state-of-the-art approaches.
- The authors release code and data publicly via GitHub, enabling replication and further research.
