Delay-Aware Diffusion Policy: Bridging the Observation-Execution Gap in Dynamic Tasks

arXiv cs.RO / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses how robot control suffers when inference delay causes a mismatch between the observed state and the state at action execution (tens to hundreds of milliseconds).
  • It proposes Delay-Aware Diffusion Policy (DA-DP), which trains and runs policies by incorporating measured delay rather than assuming zero delay.
  • DA-DP corrects zero-delay trajectories into delay-compensated versions and adds delay conditioning so the policy can adapt to different latencies.
  • Experiments across multiple tasks, robots, and delay settings show DA-DP achieves higher and more robust success rates than delay-unaware baselines.
  • The approach is architecture-agnostic and also motivates evaluation protocols that report performance versus measured latency, not only task difficulty.

Abstract

As a robot senses and selects actions, the world keeps changing. This inference delay creates a gap of tens to hundreds of milliseconds between the observed state and the state at execution. In this work, we take the natural generalization from zero delay to measured delay during training and inference. We introduce Delay-Aware Diffusion Policy (DA-DP), a framework for explicitly incorporating inference delays into policy learning. DA-DP corrects zero-delay trajectories to their delay-compensated counterparts, and augments the policy with delay conditioning. We empirically validate DA-DP on a variety of tasks, robots, and delays and find its success rate more robust to delay than delay-unaware methods. DA-DP is architecture agnostic and transfers beyond diffusion policies, offering a general pattern for delay-aware imitation learning. More broadly, DA-DP encourages evaluation protocols that report performance as a function of measured latency, not just task difficulty.