DTP-Attack: A decision-based black-box adversarial attack on trajectory prediction

arXiv cs.RO / 3/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DTP-Attack, a decision-based black-box adversarial attack framework designed to target autonomous driving trajectory prediction systems without requiring model internals or gradient access.
  • DTP-Attack relies only on binary decision outputs and uses a boundary-walking algorithm to find adversarial regions while preserving trajectory realism via proximity, avoiding rigid physical constraints.
  • The method supports two attack goals: intention misclassification (changing the perceived driving maneuver) and prediction accuracy degradation (increasing forecasting errors).
  • Experiments on nuScenes and Apolloscape using multiple state-of-the-art trajectory prediction models show strong effectiveness, including 41–81% success rates for intention misclassification with perturbations under 0.45 m.
  • The results suggest fundamental vulnerabilities in current trajectory prediction models and argue for urgent development of robust defenses in safety-critical autonomous driving contexts.

Abstract

Trajectory prediction systems are critical for autonomous vehicle safety, yet remain vulnerable to adversarial attacks that can cause catastrophic traffic behavior misinterpretations. Existing attack methods require white-box access with gradient information and rely on rigid physical constraints, limiting real-world applicability. We propose DTP-Attack, a decision-based black-box adversarial attack framework tailored for trajectory prediction systems. Our method operates exclusively on binary decision outputs without requiring model internals or gradients, making it practical for real-world scenarios. DTP-Attack employs a novel boundary walking algorithm that navigates adversarial regions without fixed constraints, naturally maintaining trajectory realism through proximity preservation. Unlike existing approaches, our method supports both intention misclassification attacks and prediction accuracy degradation. Extensive evaluation on nuScenes and Apolloscape datasets across state-of-the-art models including Trajectron++ and Grip++ demonstrates superior performance. DTP-Attack achieves 41 - 81% attack success rates for intention misclassification attacks that manipulate perceived driving maneuvers with perturbations below 0.45 m, and increases prediction errors by 1.9 - 4.2 for accuracy degradation. Our method consistently outperforms existing black-box approaches while maintaining high controllability and reliability across diverse scenarios. These results reveal fundamental vulnerabilities in current trajectory prediction systems, highlighting urgent needs for robust defenses in safety-critical autonomous driving applications.