Towards Viewpoint-Robust End-to-End Autonomous Driving with 3D Foundation Model Priors

arXiv cs.CV / 4/2/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a key limitation in end-to-end autonomous driving: many existing trajectory-planning models degrade when camera viewpoint changes from the training distribution.
  • It proposes an augmentation-free technique that uses geometric priors from a 3D foundation model by injecting per-pixel 3D positions (from depth estimates) as positional embeddings and fusing geometric intermediate features via cross-attention.
  • Experiments on the VR-Drive benchmark (camera viewpoint perturbations) show reduced performance drop across most perturbation types.
  • The approach yields the most clear improvements for pitch and height perturbations, while robustness gains for longitudinal translation are smaller, indicating a need for more viewpoint-agnostic integration.
  • Overall, the work suggests that incorporating 3D geometric priors into end-to-end pipelines can improve viewpoint robustness without relying on additional data augmentation.

Abstract

Robust trajectory planning under camera viewpoint changes is important for scalable end-to-end autonomous driving. However, existing models often depend heavily on the camera viewpoints seen during training. We investigate an augmentation-free approach that leverages geometric priors from a 3D foundation model. The method injects per-pixel 3D positions derived from depth estimates as positional embeddings and fuses intermediate geometric features through cross-attention. Experiments on the VR-Drive camera viewpoint perturbation benchmark show reduced performance degradation under most perturbation conditions, with clear improvements under pitch and height perturbations. Gains under longitudinal translation are smaller, suggesting that more viewpoint-agnostic integration is needed for robustness to camera viewpoint changes.