Routing-Free Mixture-of-Experts

arXiv cs.LG / 4/2/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes “Routing-Free Mixture-of-Experts (MoE),” removing centralized routing components (e.g., routers, softmax, top‑k, and load-balancing heuristics) in favor of fully expert-local activation.
  • It introduces a unified, adaptive load-balancing framework that optimizes both expert-usage and token-usage objectives via a configurable interpolation for more flexible resource allocation.
  • The approach is designed to be trained end-to-end using continuous gradient flow, letting each expert learn its own activation behavior without hard-coded routing biases.
  • Experiments report that Routing-Free MoE can outperform existing baselines with improved scalability and robustness, along with a detailed behavioral analysis to guide future MoE design.
  • The work aims to inform future MoE design and optimization, potentially impacting how practitioners architect and train expert-based models for efficiency and reliability.

Abstract

Standard Mixture-of-Experts (MoE) models rely on centralized routing mechanisms that introduce rigid inductive biases. We propose Routing-Free MoE which eliminates any hard-coded centralized designs including external routers, Softmax, Top-K and load balancing, instead encapsulating all activation functionalities within individual experts and directly optimized through continuous gradient flow, enabling each expert to determine its activation entirely on its own. We introduce a unified adaptive load-balancing framework to simultaneously optimize both expert-balancing and token-balancing objectives through a configurable interpolation, allowing flexible and customizable resource allocation. Extensive experiments show that Routing-Free MoE can consistently outperform baselines with better scalability and robustness. We analyze its behavior in detail and offer insights that may facilitate future MoE design ad optimization.