Stochastic Dimension-Free Zeroth-Order Estimator for High-Dimensional and High-Order PINNs

arXiv cs.LG / 3/26/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SDZE (Stochastic Dimension-Free Zeroth-Order Estimator) to train Physics-Informed Neural Networks for high-dimensional and high-order PDEs without the typical O(d^k) spatial derivative and large backpropagation memory costs.
  • It combines randomized spatial dimension-free estimation with zeroth-order (BP-free) optimization, but resolves the resulting variance blow-up (O(1/ε^2)) using Common Random Numbers Synchronization (CRNS) to keep perturbations’ randomness aligned.
  • SDZE also adds an implicit matrix-free subspace projection to cut parameter exploration variance from O(P) down to O(r) while keeping optimizer memory at O(1).
  • The authors report empirical success training extremely large PINNs (up to 10-million-dimensional) on a single NVIDIA A100 GPU, with notable speed and memory improvements over existing baselines.

Abstract

Physics-Informed Neural Networks (PINNs) for high-dimensional and high-order partial differential equations (PDEs) are primarily constrained by the \mathcal{O}(d^k) spatial derivative complexity and the \mathcal{O}(P) memory overhead of backpropagation (BP). While randomized spatial estimators successfully reduce the spatial complexity to \mathcal{O}(1), their reliance on first-order optimization still leads to prohibitive memory consumption at scale. Zeroth-order (ZO) optimization offers a BP-free alternative; however, naively combining randomized spatial operators with ZO perturbations triggers a variance explosion of \mathcal{O}(1/\varepsilon^2), leading to numerical divergence. To address these challenges, we propose the \textbf{S}tochastic \textbf{D}imension-free \textbf{Z}eroth-order \textbf{E}stimator (\textbf{SDZE}), a unified framework that achieves dimension-independent complexity in both space and memory. Specifically, SDZE leverages \emph{Common Random Numbers Synchronization (CRNS)} to algebraically cancel the \mathcal{O}(1/\varepsilon^2) variance by locking spatial random seeds across perturbations. Furthermore, an \emph{implicit matrix-free subspace projection} is introduced to reduce parameter exploration variance from \mathcal{O}(P) to \mathcal{O}(r) while maintaining an \mathcal{O}(1) optimizer memory footprint. Empirical results demonstrate that SDZE enables the training of 10-million-dimensional PINNs on a single NVIDIA A100 GPU, delivering significant improvements in speed and memory efficiency over state-of-the-art baselines.