Mixture of Experts Framework in Machine Learning Interatomic Potentials for Atomistic Simulations

arXiv cs.LG / 4/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles the high inference cost of ML interatomic potentials by proposing a multifidelity Mixture-of-Experts framework for large-scale atomistic simulations.
  • It partitions the simulation space into chemically complex and simple regions, assigning different-capacity E(3)-equivariant Allegro-based models to each region.
  • A key challenge—mechanical mismatch at expert-model interfaces that can cause artificial stresses and instability—is addressed via a co-training strategy with agreement constraints on per-atom energy and force in shared bulk environments.
  • Experiments on a Pt+CO catalytic system show the co-trained experts preserve exact energy conservation, match bulk mechanical properties, and reach accuracy comparable to full high-fidelity simulation while running at more than twice the computational speed.

Abstract

First-principles atomistic simulations are essential for understanding complex material phenomena but are fundamentally limited by their computational cost. While Machine Learning Interatomic Potentials (MLIPs) have drastically improved cost for a given accuracy, their inference cost remains a bottleneck for massive systems or long timescales. To address this, we introduce a multifidelity "Mixture-of-Experts" framework based on the E(3)-equivariant Allegro architecture. Our method spatially partitions the simulation domain into a chemically complex region (e.g., reactive interfaces) and a simple region (e.g., bulk lattice), assigning models of varying capacity to each. Among the challenges in such static domain decomposition, the mechanical mismatch between models at the interface is particularly critical, as it can generate artificial stress fields and instability. We address this challenge with a co-training strategy in which the loss function includes agreement constraints -- penalties on per-atom energy and force discrepancies between models evaluated on shared bulk environments -- forcing the independent models to learn a consistent physical description of the bulk material. We validate this approach on a realistic Pt+CO catalytic system, demonstrating that the co-trained models maintain exact energy conservation, align their bulk mechanical response (e.g., equation of state and bulk modulus), and achieve predictive accuracy comparable to a full high-fidelity simulation at more than twice the computational speed.