Design and Behavior of Sparse Mixture-of-Experts Layers in CNN-based Semantic Segmentation

arXiv cs.CV / 4/16/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how to integrate sparse mixture-of-experts (MoE) layers into CNN-based semantic segmentation using a coarser, patch-wise routing strategy rather than fine-grained filter/channel MoE designs.
  • Experiments on Cityscapes and BDD100K (with encoder-decoder and backbone-based CNNs) analyze how architectural choices influence routing dynamics and expert specialization.
  • Results show consistent, architecture-dependent segmentation quality gains of up to +3.9 mIoU with little added computational overhead.
  • The authors find strong sensitivity to design decisions, indicating that sparse MoE performance in CNN dense prediction depends heavily on layer/routing configuration.
  • The paper provides empirical insights and publishes code at GitHub to support further experimentation with MoE layers in CNN segmentation pipelines.

Abstract

Sparse mixture-of-experts (MoE) layers have been shown to substantially increase model capacity without a proportional increase in computational cost and are widely used in transformer architectures, where they typically replace feed-forward network blocks. In contrast, integrating sparse MoE layers into convolutional neural networks (CNNs) remains inconsistent, with most prior work focusing on fine-grained MoEs operating at the filter or channel levels. In this work, we investigate a coarser, patch-wise formulation of sparse MoE layers for semantic segmentation, where local regions are routed to a small subset of convolutional experts. Through experiments on the Cityscapes and BDD100K datasets using encoder-decoder and backbone-based CNNs, we conduct a design analysis to assess how architectural choices affect routing dynamics and expert specialization. Our results demonstrate consistent, architecture-dependent improvements (up to +3.9 mIoU) with little computational overhead, while revealing strong design sensitivity. Our work provides empirical insights into the design and internal dynamics of sparse MoE layers in CNN-based dense prediction. Our code is available at https://github.com/KASTEL-MobilityLab/moe-layers/.