Multi-Domain Learning with Global Expert Mapping
arXiv cs.CV / 4/22/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses a common limitation of vision models: they often do not generalize well to new domains beyond their training distributions, motivating multi-dataset learning for robustness to domain shift.
- It argues that existing Mixture-of-Experts (MoE) approaches can under-specialize because load-balancing encourages uniform routing, which conflicts with domain-aware specialization and hurts performance on rare or out-of-distribution domains.
- The authors propose GEM (Global Expert Mapping), a planner–compiler framework that replaces the learned router with a global scheduler that computes dataset-to-expert assignments via linear programming relaxation.
- A hierarchical rounding compiler then converts the fractional plan into a deterministic, capacity-aware routing map, aiming to avoid balancing loss and produce more interpretable routing behavior.
- Experiments report that GEM-DINO reaches state-of-the-art results on the UODB benchmark, including gains on underrepresented datasets and improved behavior in few-shot adaptation by reducing task interference.


