Rectify, Don't Regret: Avoiding Pitfalls of Differentiable Simulation in Trajectory Prediction

arXiv cs.RO / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that fully differentiable closed-loop trajectory simulators can suffer from shortcut learning, where backpropagated gradients leak future ground-truth information through induced state inputs.
  • It proposes a “detached receding horizon rollout” method that explicitly breaks the computation graph across simulation steps to prevent non-causal “regret-based” optimization of past predictions.
  • Experiments on nuScenes and DeepScenario show the approach improves robustness of recovery from drifted states, reducing target collisions by up to 33.24% versus fully differentiable closed-loop training at high replanning frequencies.
  • Compared with standard open-loop baselines, the non-differentiable training framework also reduces collisions by up to 27.74% in dense environments while improving multi-modal prediction diversity and lane alignment.

Abstract

Current open-loop trajectory models struggle in real-world autonomous driving because minor initial deviations often cascade into compounding errors, pushing the agent into out-of-distribution states. While fully differentiable closed-loop simulators attempt to address this, they suffer from shortcut learning: the loss gradients flow backward through induced state inputs, inadvertently leaking future ground truth information directly into the model's own previous predictions. The model exploits these signals to artificially avoid drift, non-causally "regretting" past mistakes rather than learning genuinely reactive recovery. To address this, we introduce a detached receding horizon rollout. By explicitly severing the computation graph between simulation steps, the model learns genuine recovery behaviors from drifted states, forcing it to "rectify" mistakes rather than non-causally optimizing past predictions. Extensive evaluations on the nuScenes and DeepScenario datasets show our approach yields more robust recovery strategies, reducing target collisions by up to 33.24% compared to fully differentiable closed-loop training at high replanning frequencies. Furthermore, compared to standard open-loop baselines, our non-differentiable framework decreases collisions by up to 27.74% in dense environments while simultaneously improving multi-modal prediction diversity and lane alignment.