Late Fusion Neural Operators for Extrapolation Across Parameter Space in Partial Differential Equations

arXiv cs.LG / 4/21/2026

📰 NewsModels & Research

Key Points

  • The paper addresses a core challenge for neural operators: accurately extrapolating PDE solutions to parameter regimes not seen during training despite distribution shifts caused by varying physical parameters.
  • It proposes the “Late Fusion Neural Operator,” designed to disentangle state dynamics learning from parameter effects when state and parameter representations are entangled.
  • The method uses neural operators to learn latent state representations and incorporates parameter information via sparse regression in a structured way.
  • Experiments on four PDE benchmarks (including advection, Burgers, and 1D/2D reaction-diffusion) show consistent improvements over Fourier Neural Operator and CAPE-FNO.
  • Late Fusion Neural Operators achieve the best results overall, reducing RMSE by an average of 72.9% in-domain and 71.8% out-domain versus the second-best approach, demonstrating strong generalization.

Abstract

Developing neural operators that accurately predict the behavior of systems governed by partial differential equations (PDEs) across unseen parameter regimes is crucial for robust generalization in scientific and engineering applications. In practical applications, variations in physical parameters induce distribution shifts between training and prediction regimes, making extrapolation a central challenge. As a result, the way parameters are incorporated into neural operator models plays a key role in their ability to generalize, particularly when state and parameter representations are entangled. In this work, we introduce the Late Fusion Neural Operator, an architecture that disentangles learning state dynamics from parameter effects, improving predictive performance both within and beyond the training distribution. Our approach combines neural operators for learning latent state representations with sparse regression to incorporate parameter information in a structured manner. Across four benchmark PDEs including advection, Burgers, and both 1D and 2D reaction-diffusion equations, the proposed method consistently outperforms Fourier Neural Operator and CAPE-FNO. Late Fusion Neural Operators achieve consistently the best performance in all experiments, with an average RMSE reduction of 72.9% in-domain and 71.8% out-domain compared to the second-best method. These results demonstrate strong generalization across both in-domain and out-domain parameter regimes.

Late Fusion Neural Operators for Extrapolation Across Parameter Space in Partial Differential Equations | AI Navigate