AI Navigate

Optimal Expert-Attention Allocation in Mixture-of-Experts: A Scalable Law for Dynamic Model Design

arXiv cs.LG / 3/12/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The authors extend neural scaling laws to Mixture-of-Experts (MoE) models and study how to allocate compute between expert and attention sub-layers.
  • They define the ratio r as the fraction of FLOPs per token spent on expert layers versus attention layers and analyze its interaction with total compute and model sparsity.
  • Through experiments with GPT-style MoE Transformers, they find that the optimal ratio r* follows a power-law with total compute and varies with sparsity.
  • They derive an explicit formula for r*, enabling precise control over expert-attention compute allocation and generalizing the Chinchilla scaling law to incorporate this architectural parameter.
  • The work provides practical guidelines for designing efficient MoE models that maximize performance under fixed compute budgets.

Abstract

This paper presents a novel extension of neural scaling laws to Mixture-of-Experts (MoE) models, focusing on the optimal allocation of compute between expert and attention sub-layers. As MoE architectures have emerged as an efficient method for scaling model capacity without proportionally increasing computation, determining the optimal expert-attention compute ratio becomes critical. We define the ratio r as the fraction of total FLOPs per token dedicated to the expert layers versus the attention layers, and explore how this ratio interacts with the overall compute budget and model sparsity. Through extensive experiments with GPT-style MoE Transformers, we empirically find that the optimal ratio r^* follows a power-law relationship with total compute and varies with sparsity. Our analysis leads to an explicit formula for r^*, enabling precise control over the expert-attention compute allocation. We generalize the Chinchilla scaling law by incorporating this architectural parameter, providing a new framework for tuning MoE models beyond size and data. Our findings offer practical guidelines for designing efficient MoE models, optimizing performance while respecting fixed compute budgets.