AI Navigate

Manifold-Optimal Guidance: A Unified Riemannian Control View of Diffusion Guidance

arXiv cs.CV / 3/13/2026

📰 NewsModels & Research

Key Points

  • MOG identifies a geometric mismatch in classifier-free guidance that can cause sampling trajectories to drift off the data manifold.
  • It reframes diffusion guidance as a local optimal control problem and delivers a closed-form, geometry-aware Riemannian update that corrects off-manifold drift without retraining.
  • Auto-MOG introduces a dynamic energy-balancing schedule that adaptively calibrates guidance strength and eliminates the need for manual hyperparameter tuning.
  • The approach requires no retraining and imposes virtually no additional computational overhead.
  • Extensive validation demonstrates superior fidelity and alignment relative to baselines.

Abstract

Classifier-Free Guidance (CFG) serves as the de facto control mechanism for conditional diffusion, yet high guidance scales notoriously induce oversaturation, texture artifacts, and structural collapse. We attribute this failure to a geometric mismatch: standard CFG performs Euclidean extrapolation in ambient space, inadvertently driving sampling trajectories off the high-density data manifold. To resolve this, we present Manifold-Optimal Guidance (MOG), a framework that reformulates guidance as a local optimal control problem. MOG yields a closed-form, geometry-aware Riemannian update that corrects off-manifold drift without requiring retraining. Leveraging this perspective, we further introduce Auto-MOG, a dynamic energy-balancing schedule that adaptively calibrates guidance strength, effectively eliminating the need for manual hyperparameter tuning. Extensive validation demonstrates that MOG yields superior fidelity and alignment compared to baselines, with virtually no added computational overhead.