Curvature-Aware Optimization for High-Accuracy Physics-Informed Neural Networks

arXiv cs.LG / 4/8/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes curvature-aware optimization techniques to speed up and stabilize training of physics-informed neural networks (PINNs) for both PDEs and ODEs.
  • It introduces efficient implementations of Natural Gradient, Self-Scaling BFGS, and Broyden-type optimizers, targeting faster convergence to high accuracy on hard physical problems.
  • Experiments cover benchmark equations such as the Helmholtz equation, Stokes flow, inviscid Burgers, Euler equations for high-speed flows, and stiff pharmacokinetics/pharmacodynamics ODEs.
  • Beyond optimizer work, the authors present new PINN-based solution methods for the inviscid Burgers and Euler equations and validate results against high-order numerical solvers.
  • The study also discusses how to scale quasi-Newton optimizers for batched training to support larger, data-driven scientific machine learning workflows.

Abstract

Efficient and robust optimization is essential for neural networks, enabling scientific machine learning models to converge rapidly to very high accuracy -- faithfully capturing complex physical behavior governed by differential equations. In this work, we present advanced optimization strategies to accelerate the convergence of physics-informed neural networks (PINNs) for challenging partial (PDEs) and ordinary differential equations (ODEs). Specifically, we provide efficient implementations of the Natural Gradient (NG) optimizer, Self-Scaling BFGS and Broyden optimizers, and demonstrate their performance on problems including the Helmholtz equation, Stokes flow, inviscid Burgers equation, Euler equations for high-speed flows, and stiff ODEs arising in pharmacokinetics and pharmacodynamics. Beyond optimizer development, we also propose new PINN-based methods for solving the inviscid Burgers and Euler equations, and compare the resulting solutions against high-order numerical methods to provide a rigorous and fair assessment. Finally, we address the challenge of scaling these quasi-Newton optimizers for batched training, enabling efficient and scalable solutions for large data-driven problems.