MoE-GRPO: Optimizing Mixture-of-Experts via Reinforcement Learning in Vision-Language Models

arXiv cs.CV / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • Mixture-of-Experts (MoE) reduces Transformer compute by sparsely activating experts per token, and this idea has been extended to Vision-Language Models (VLMs) to improve multimodal scalability.
  • The paper argues that deterministic top-K expert routing can miss better expert combinations and cause expert overfitting due to insufficient routing diversity.
  • MoE-GRPO proposes an RL framework that treats expert selection as sequential decision-making and optimizes routing with Group Relative Policy Optimization (GRPO) to learn adaptive routing policies.
  • It also introduces a modality-aware router guidance mechanism to stabilize and speed up training by discouraging exploration of rarely-used experts for a given modality (e.g., image vs. video).
  • Experiments on multimodal image and video benchmarks show MoE-GRPO outperforms standard top-K routing and variants by improving expert diversity and enabling task-level expert specialization while mitigating overfitting.

Abstract

Mixture-of-Experts (MoE) has emerged as an effective approach to reduce the computational overhead of Transformer architectures by sparsely activating a subset of parameters for each token while preserving high model capacity. This paradigm has recently been extended to Vision-Language Models (VLMs), enabling scalable multi-modal understanding with reduced computational cost. However, the widely adopted deterministic top-K routing mechanism may overlook more optimal expert combinations and lead to expert overfitting. To address this limitation and improve the diversity of expert selection, we propose MoE-GRPO, a reinforcement learning (RL)-based framework for optimizing expert routing in MoE-based VLMs. Specifically, we formulate expert selection as a sequential decision-making problem and optimize it using Group Relative Policy Optimization (GRPO), allowing the model to learn adaptive expert routing policies through exploration and reward-based feedback. Furthermore, we introduce a modality-aware router guidance that enhances training stability and efficiency by discouraging the router from exploring experts that are infrequently activated for a given modality. Extensive experiments on multi-modal image and video benchmarks show that MoE-GRPO consistently outperforms standard top-K routing and its variants by promoting more diverse expert selection, thereby mitigating expert overfitting and enabling a task-level expert specialization.