Factored Levenberg-Marquardt for Diffeomorphic Image Registration: An efficient optimizer for FireANTs

arXiv cs.CV / 3/23/2026

📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • FireANTs introduces a modified Levenberg–Marquardt optimizer with a single scalar damping parameter, adaptively tuned via a trust-region approach, for diffeomorphic image registration.
  • The new optimizer reduces memory usage by up to 24.6% for large volumes while preserving performance across four datasets.
  • A single hyperparameter configuration tuned on brain MRI transfers to lung CT and cross-modal abdominal registration, matching or outperforming Adam on three of four benchmarks.
  • Ablation experiments show that a Metropolis-Hastings style rejection step can prevent updates that worsen the loss function.

Abstract

FireANTs introduced a novel Eulerian descent method for plug-and-play behavior with arbitrary optimizers adapted for diffeomorphic image registration as a test-time optimization problem, with a GPU-accelerated implementation. FireANTs uses Adam as its default optimizer for fast and more robust optimization. However, Adam requires storing state variables (i.e. momentum and squared-momentum estimates), each of which can consume significant memory, prohibiting its use for significantly large images. In this work, we propose a modified Levenberg-Marquardt (LM) optimizer that requires only a single scalar damping parameter as optimizer state, that is adaptively tuned using a trust region approach. The resulting optimizer reduces memory by up to 24.6% for large volumes, and retaining performance across all four datasets. A single hyperparameter configuration tuned on brain MRI transfers without modification to lung CT and cross-modal abdominal registration, matching or outperforming Adam on three of four benchmarks. We also perform ablations on the effectiveness of using Metropolis-Hastings style rejection step to prevent updates that worsen the loss function.