MILE: Mixture of Incremental LoRA Experts for Continual Semantic Segmentation across Domains and Modalities

arXiv cs.CV / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes MILE, a modular, parameter-efficient approach for continual semantic segmentation that learns new domains or modalities without degrading past performance.
  • MILE uses LoRA to add lightweight expert modules per new task while keeping the pretrained base network frozen, reducing catastrophic forgetting.
  • A prototype-guided gating mechanism dynamically chooses the best expert during inference for improved task/domain/modal alignment.
  • The authors report that MILE delivers strong benchmark results with better stability, plasticity, and scalability, using only a small parameter increase per task and far fewer storage costs than full model expansions.

Abstract

Continual semantic segmentation requires models to adapt to new domains or modalities without sacrificing performance on previously learned tasks. Expert-based learning, in which task-specific modules specialize in different domains, has proven effective in mitigating forgetting. These methods include dynamic expansion, which suffers from scalability issues, or parameter isolation, which constrains the ability to learn new tasks. We introduce Mixture of Incremental LoRA Experts (MILE), a modular and parameter-efficient framework for continual segmentation across both domains and modalities. MILE leverages Low-Rank Adaptation (LoRA) to instantiate lightweight experts for each new task while keeping the pretrained base network frozen. Each expert is trained exclusively on its task data, thus avoids overwriting previously learned information. A prototype-guided gating mechanism dynamically selects the most appropriate expert at inference. MILE achieves the benefits of expert-based learning while overcoming its scalability limitations. It requires only a marginal parameter increase per task and tens of LoRA adapters are needed before matching the size of a single full model, making it highly efficient in both training and storage. Across domain- and modality-incremental benchmarks, MILE achieves strong performance while ensuring better stability, plasticity, and scalability.