Transferable SCF-Acceleration through Solver-Aligned Initialization Learning

arXiv cs.LG / 4/24/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies why ML-predicted initial guesses for SCF calculations can hurt performance on larger molecules: the issue is largely due to a supervision mismatch rather than pure extrapolation beyond the training distribution.
  • It proposes Solver-Aligned Initialization Learning (SAIL), which trains initial-guess models by differentiating end-to-end through the SCF solver for both Hamiltonian and density-matrix formulations.
  • The authors introduce the Effective Relative Iteration Count (ERIC), a revised metric that corrects the commonly used RIC by accounting for hidden overhead such as Fock-build cost.
  • Experiments on QM40 (up to ~4× larger than training) show substantial convergence improvements, with ERIC reduced by 37% (PBE), 33% (SCAN), and 27% (B3LYP), exceeding prior SOTA gains.
  • On QMugs (about ~10× larger than training), SAIL achieves about a 1.25× wall-time speedup at the hybrid level of theory, extending ML-accelerated SCF to larger, drug-like molecules.

Abstract

Machine learning methods that predict initial guesses from molecular geometry can reduce this cost, but matrix-prediction models fail when extrapolating to larger molecules, degrading rather than accelerating convergence [Liu et al. 2025]. We show that this failure is a supervision problem, not an extrapolation problem: models trained on ground-state targets fit those targets well out of distribution, yet produce initial guesses that slow convergence. Solver-Aligned Initialization Learning (SAIL) resolves this for both Hamiltonian and density matrix models by differentiating through the SCF solver end-to-end. We introduce the Effective Relative Iteration Count (ERIC), a correction to the commonly used RIC that accounts for hidden Fock-build overhead. On QM40, containing molecules up to 4\times larger than the training distribution, SAIL reduces ERIC by 37% (PBE), 33% (SCAN), and 27% (B3LYP), more than doubling the previous state-of-the-art reduction on B3LYP (10%). On QMugs molecules 10\times the training size, SAIL delivers a 1.25\times wall-time speedup at the hybrid level of theory, extending ML SCF acceleration to large drug-like molecules.