Symbiotic-MoE: Unlocking the Synergy between Generation and Understanding

arXiv cs.CL / 4/10/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Large multimodal models that learn image generation can suffer catastrophic forgetting in understanding tasks due to severe gradient conflicts, motivating approaches beyond existing mixture architectures.
  • The paper introduces Symbiotic-MoE, a unified pre-training framework that keeps a native multimodal MoE Transformer structure while preventing task interference with zero-parameter overhead.
  • It identifies a failure mode in standard MoE tuning—routing collapse—where generative gradients dominate expert utilization, and addresses it via modality-aware expert disentanglement that uses shared experts as a semantic bridge.
  • A progressive training strategy with differential learning rates and early-stage gradient shielding is proposed to protect pretrained knowledge early on, then turn generative signals into constructive feedback for understanding.
  • Experiments report faster generative convergence and improved cross-modal synergy, with gains on benchmarks including MMLU and OCRBench.

Abstract

Empowering Large Multimodal Models (LMMs) with image generation often leads to catastrophic forgetting in understanding tasks due to severe gradient conflicts. While existing paradigms like Mixture-of-Transformers (MoT) mitigate this conflict through structural isolation, they fundamentally sever cross-modal synergy and suffer from capacity fragmentation. In this work, we present Symbiotic-MoE, a unified pre-training framework that resolves task interference within a native multimodal Mixture-of-Experts (MoE) Transformers architecture with zero-parameter overhead. We first identify that standard MoE tuning leads to routing collapse, where generative gradients dominate expert utilization. To address this, we introduce Modality-Aware Expert Disentanglement, which partitions experts into task-specific groups while utilizing shared experts as a multimodal semantic bridge. Crucially, this design allows shared experts to absorb fine-grained visual semantics from generative tasks to enrich textual representations. To optimize this, we propose a Progressive Training Strategy featuring differential learning rates and early-stage gradient shielding. This mechanism not only shields pre-trained knowledge from early volatility but eventually transforms generative signals into constructive feedback for understanding. Extensive experiments demonstrate that Symbiotic-MoE achieves rapid generative convergence while unlocking cross-modal synergy, boosting inherent understanding with remarkable gains on MMLU and OCRBench.