Algorithmic warm starts for Hamiltonian Monte Carlo

arXiv stat.ML / 2026/3/25

💬 オピニオンIdeas & Deep AnalysisModels & Research

要点

  • The paper tackles a key open question about Hamiltonian Monte Carlo (HMC) in high dimensions: how many iterations are needed to reach a “warm start” close to stationarity as a function of dimension d.
  • It shows that while Metropolized HMC can converge in ~O(d^{1/4}) from a warm start, it is substantially slower without one, motivating warm-start finding as the main computational bottleneck.
  • Under assumptions of strong log-concavity (or isoperimetry) and bounded third-order derivatives, the authors prove that non-Metropolized HMC can produce a warm start in ~O(d^{1/4}) iterations.
  • After obtaining this warm start, the method switches to Metropolized HMC to achieve high-accuracy sampling with an overall ~O(d^{1/4}) complexity, improving on a previous best of ~O(d^{1/2}).
  • The work provides both a dimensional-complexity resolution for MH/HMC theory in this regime and a practical warm-start prescription for implementations.

Abstract

Generating samples from a continuous probability density is a central algorithmic problem across statistics, engineering, and the sciences. For high-dimensional settings, Hamiltonian Monte Carlo (HMC) is the default algorithm across mainstream software packages. However, despite the extensive line of work on HMC and its widespread empirical success, it remains unclear how many iterations of HMC are required as a function of the dimension d. On one hand, a variety of results show that Metropolized HMC converges in O(d^{1/4}) iterations from a warm start close to stationarity. On the other hand, Metropolized HMC is significantly slower without a warm start, e.g., requiring \Omega(d^{1/2}) iterations even for simple target distributions such as isotropic Gaussians. Finding a warm start is therefore the computational bottleneck for HMC. We resolve this issue for the well-studied setting of sampling from a probability distribution satisfying strong log-concavity (or isoperimetry) and third-order derivative bounds. We prove that \emph{non-Metropolized} HMC generates a warm start in \tilde{O}(d^{1/4}) iterations, after which we can exploit the warm start using Metropolized HMC. Our final complexity of \tilde{O}(d^{1/4}) is the fastest algorithm for high-accuracy sampling under these assumptions, improving over the prior best of \tilde{O}(d^{1/2}). This closes the long line of work on the dimensional complexity of MHMC for such settings, and also provides a simple warm-start prescription for practical implementations.