FODMP: Fast One-Step Diffusion of Movement Primitives Generation for Time-Dependent Robot Actions

arXiv cs.AI / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a limitation of diffusion-based robot learning: existing action-chunking diffusion policies are fast but only generate short, reactive motion segments, missing time-dependent movement primitives.
  • It builds on Movement Primitive Diffusion (MPD), which uses ProDMPs to represent temporally structured trajectories, but MPD remains too slow because the motion decoder is embedded in a multi-step diffusion process.
  • The authors propose FODMP, which distills diffusion models into the ProDMP trajectory parameter space and generates motions with a single-step decoder to remove the inference bottleneck.
  • Experiments on MetaWorld and ManiSkill show FODMP can run up to 10× faster than MPD and 7× faster than action-chunking diffusion policies while maintaining or improving success rates.
  • The framework also enables dynamic acceleration–deceleration primitives that improve real-time tasks such as intercepting and catching a fast-flying ball under closed-loop vision control.

Abstract

Diffusion models are increasingly used for robot learning, but current designs face a clear trade-off. Action-chunking diffusion policies like ManiCM are fast to run, yet they only predict short segments of motion. This makes them reactive, but unable to capture time-dependent motion primitives, such as following a spring-damper-like behavior with built-in dynamic profiles of acceleration and deceleration. Recently, Movement Primitive Diffusion (MPD) partially addresses this limitation by parameterizing full trajectories using Probabilistic Dynamic Movement Primitives (ProDMPs), thereby enabling the generation of temporally structured motions. Nevertheless, MPD integrates the motion decoder directly into a multi-step diffusion process, resulting in prohibitively high inference latency that limits its applicability in real-time control settings. We propose FODMP (Fast One-step Diffusion of Movement Primitives), a new framework that distills diffusion models into the ProDMPs trajectory parameter space and generates motion using a single-step decoder. FODMP retains the temporal structure of movement primitives while eliminating the inference bottleneck through single-step consistency distillation. This enables robots to execute time-dependent primitives at high inference speed, suitable for closed-loop vision-based control. On standard manipulation benchmarks (MetaWorld, ManiSkill), FODMP runs up to 10 times faster than MPD and 7 times faster than action-chunking diffusion policies, while matching or exceeding their success rates. Beyond speed, by generating fast acceleration-deceleration motion primitives, FODMP allows the robot to intercept and securely catch a fast-flying ball, whereas action-chunking diffusion policy and MPD respond too slowly for real-time interception.
広告