Bayesian Optimization with Gaussian Processes to Accelerate Stationary Point Searches

arXiv stat.ML / 4/17/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper shows that building surrogate models can reduce the number of expensive electronic-structure (potential energy surface) evaluations for stationary point searches by about an order of magnitude without sacrificing the accuracy of the underlying theory.
  • It proposes a unified Bayesian optimization framework that covers three tasks—energy minimization, single-point saddle searches, and double-ended path searches—using the same six-step surrogate loop with differences only in the inner optimization target and acquisition criterion.
  • The method relies on Gaussian process regression with derivative observations, inverse-distance kernels, and active learning to more efficiently decide where to evaluate next.
  • It introduces optional production-oriented extensions for scalability and robustness, including farthest-point sampling via Earth Mover’s Distance, MAP regularization, an adaptive trust radius, and random Fourier features.
  • To support practical adoption, the authors provide educational Rust code demonstrating that all three applications share the same Bayesian optimization loop, connecting theory to implementation.

Abstract

Building local surrogates to accelerate stationary point searches on potential energy surfaces spans decades of effort. Done correctly, surrogates can reduce the number of expensive electronic structure evaluations by roughly an order of magnitude while preserving the accuracy of the underlying theory, with the gain depending on oracle cost, search distance, and the availability of analytical forces. We present a unified Bayesian optimization view of minimization, single-point saddle searches, and double-ended path searches: all three share one six-step surrogate loop and differ only in the inner optimization target and the acquisition criterion. The framework uses Gaussian process regression with derivative observations, inverse-distance kernels, and active learning, and we develop optional extensions for production use, including farthest-point sampling with the Earth Mover's Distance, MAP regularization, an adaptive trust radius, and random Fourier features for scaling. Accompanying pedagogical Rust code demonstrates that all three applications use the same Bayesian optimization loop, bridging the gap between theoretical formulation and practical execution.