AI Navigate

Riemannian MeanFlow for One-Step Generation on Manifolds

arXiv cs.LG / 3/12/2026

📰 NewsModels & Research

Key Points

  • We propose Riemannian MeanFlow (RMF), which extends MeanFlow to manifold-valued generation by accommodating velocities in location-dependent tangent spaces.
  • RMF defines an average-velocity field via parallel transport and derives a Riemannian MeanFlow identity that links average and instantaneous velocities for intrinsic supervision.
  • To enable practical optimization, the RMF objective is decomposed into two terms and trained with conflict-aware multi-task learning to mitigate gradient interference.
  • The framework supports conditional generation via classifier-free guidance and achieves competitive one-step sampling on spheres, tori, and SO(3) with improved quality-efficiency and reduced sampling cost.

Abstract

Flow Matching enables simulation-free training of generative models on Riemannian manifolds, yet sampling typically still relies on numerically integrating a probability-flow ODE. We propose Riemannian MeanFlow (RMF), extending MeanFlow to manifold-valued generation where velocities lie in location-dependent tangent spaces. RMF defines an average-velocity field via parallel transport and derives a Riemannian MeanFlow identity that links average and instantaneous velocities for intrinsic supervision. We make this identity practical in a log-map tangent representation, avoiding trajectory simulation and heavy geometric computations. For stable optimization, we decompose the RMF objective into two terms and apply conflict-aware multi-task learning to mitigate gradient interference. RMF also supports conditional generation via classifier-free guidance. Experiments on spheres, tori, and SO(3) demonstrate competitive one-step sampling with improved quality-efficiency trade-offs and substantially reduced sampling cost.