Energy-Weighted Flow Matching: Unlocking Continuous Normalizing Flows for Efficient and Scalable Boltzmann Sampling

arXiv stat.ML / 4/22/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Energy-Weighted Flow Matching (EWFM), a new training objective for continuous normalizing flows to sample from Boltzmann (unnormalized) target distributions using only energy evaluations.
  • EWFM reformulates conditional flow matching with importance sampling, enabling training from samples drawn from arbitrary proposal distributions rather than requiring large datasets from the target distribution.
  • The authors propose two variants—iterative EWFM (iEWFM) to progressively refine proposal distributions and annealed EWFM (aEWFM) that adds temperature annealing to handle difficult energy landscapes.
  • On benchmarks such as 55-particle Lennard-Jones clusters, the method achieves sample quality competitive with established energy-only approaches while cutting energy-evaluation cost by up to about three orders of magnitude.

Abstract

Sampling from unnormalized target distributions, e.g.\ Boltzmann distributions \mu_{\text{target}}(x) \propto \exp(-E(x)/T), is fundamental to many scientific applications yet computationally challenging due to complex, high-dimensional energy landscapes. Existing approaches applying modern generative models to Boltzmann distributions either require large datasets of samples drawn from the target distribution or, when using only energy evaluations for training, cannot efficiently leverage the expressivity of advanced architectures like continuous normalizing flows that have shown promise for molecular sampling. To address these shortcomings, we introduce Energy-Weighted Flow Matching (EWFM), a novel training objective enabling continuous normalizing flows to model Boltzmann distributions using only energy function evaluations. Our objective reformulates conditional flow matching via importance sampling, allowing training with samples from arbitrary proposal distributions. Based on this objective, we develop two algorithms: iterative EWFM (iEWFM), which progressively refines proposals through iterative training, and annealed EWFM (aEWFM), which additionally incorporates temperature annealing for challenging energy landscapes. On benchmark systems, including challenging 55-particle Lennard-Jones clusters, our algorithms demonstrate sample quality competitive with established energy-only methods while requiring up to three orders of magnitude fewer energy evaluations.