Performance of Neural and Polynomial Operator Surrogates

arXiv cs.LG / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies surrogate operators for expensive parameter-to-solution maps in parametric PDEs, comparing neural operator surrogates with polynomial reduced-basis surrogates under varying input regularity.
  • It evaluates reduced-basis neural operators (trained with L2_μ and H1_μ objectives) and Fourier neural operators against reduced-basis sparse-grid and reduced-basis tensor-train polynomial surrogates on both diffusion (linear) and hyperelasticity (nonlinear) problems.
  • For smooth inputs (spectral decay rate s ≥ 2), polynomial surrogates are substantially more data efficient, and the sparse-grid method’s convergence matches theoretical expectations.
  • For rough inputs (s ≤ 1), the Fourier neural operator achieves the fastest convergence rates, indicating neural operators can be better suited to lower regularity.
  • The authors find that derivative-informed (e.g., Jacobian-based) training improves data efficiency versus standard L2_μ training, offering a competitive low-data approach when derivative information is available at reasonable cost.

Abstract

We consider the problem of constructing surrogate operators for parameter-to-solution maps arising from parametric partial differential equations, where repeated forward model evaluations are computationally expensive. We present a systematic empirical comparison of neural operator surrogates, including a reduced-basis neural operator trained with L^2_\mu and H^1_\mu objectives and the Fourier neural operator, against polynomial surrogate methods, specifically a reduced-basis sparse-grid surrogate and a reduced-basis tensor-train surrogate. All methods are evaluated on a linear parametric diffusion problem and a nonlinear parametric hyperelasticity problem, using input fields with algebraically decaying spectral coefficients at varying rates of decay s. To enable fair comparisons, we analyze ensembles of surrogate models generated by varying hyperparameters and compare the resulting Pareto frontiers of cost versus approximation accuracy, decomposing cost into contributions from data generation, setup, and evaluation. Our results show that no single method is universally superior. Polynomial surrogates achieve substantially better data efficiency for smooth input fields (s \geq 2), with convergence rates for the sparse-grid surrogate in agreement with theoretical predictions. For rough inputs (s \leq 1), the Fourier neural operator displays the fastest convergence rates. Derivative-informed training consistently improves data efficiency over standard L^2_\mu training, providing a competitive alternative for rough inputs in the low-data regime when Jacobian information is available at reasonable cost. These findings highlight the importance of matching the surrogate methodology to the regularity of the problem as well as accuracy demands and computational constraints of the application.

Performance of Neural and Polynomial Operator Surrogates | AI Navigate