MolDA: Molecular Understanding and Generation via Large Language Diffusion Model

arXiv cs.AI / 4/7/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current multimodal molecular models often use autoregressive (left-to-right) backbones that struggle with global chemical constraints like ring closures and can accumulate structural errors during generation.
  • It introduces MolDA, a multimodal molecular framework that replaces the autoregressive backbone with a discrete large language diffusion model paired with a hybrid graph encoder and a Q-Former to map structures into a language-token space.
  • The authors reformulate “Molecular Structure Preference Optimization” for masked diffusion and emphasize bidirectional iterative denoising to improve global coherence and chemical validity.
  • MolDA is presented as supporting multiple tasks—molecule generation, captioning, and property prediction—while aiming for robust reasoning enabled by the diffusion-based formulation.
  • The work is positioned as a research contribution (arXiv announcement) advancing model architecture choices for chemically valid molecule synthesis beyond AR inductive bias.

Abstract

Large Language Models (LLMs) have significantly advanced molecular discovery, but existing multimodal molecular architectures fundamentally rely on autoregressive (AR) backbones. This strict left-to-right inductive bias is sub-optimal for generating chemically valid molecules, as it struggles to account for non-local global constraints (e.g., ring closures) and often accumulates structural errors during sequential generation. To address these limitations, we propose MolDA (Molecular language model with masked Diffusion with mAsking), a novel multimodal framework that replaces the conventional AR backbone with a discrete Large Language Diffusion Model. MolDA extracts comprehensive structural representations using a hybrid graph encoder, which captures both local and global topologies, and aligns them into the language token space via a Q-Former. Furthermore, we mathematically reformulate Molecular Structure Preference Optimization specifically for the masked diffusion. Through bidirectional iterative denoising, MolDA ensures global structural coherence, chemical validity, and robust reasoning across molecule generation, captioning, and property prediction.