Pseudo-Expert Regularized Offline RL for End-to-End Autonomous Driving in Photorealistic Closed-Loop Environments

arXiv cs.RO / 4/10/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a camera-only end-to-end offline RL approach for autonomous driving that trains from a fixed simulator dataset without additional exploration, aiming to avoid imitation-learning failure modes.
  • To reduce offline RL instability from out-of-distribution action overestimation, the method regularizes training using pseudo ground-truth trajectories derived from expert logs.
  • Experiments are performed in a neural rendering closed-loop environment learned from the public nuScenes dataset, focusing on driving safety and efficiency metrics.
  • The authors report substantial improvements over imitation learning baselines, including lower collision rates and higher route completion.
  • An open-source implementation is provided, enabling others to reproduce and build on the proposed pseudo-expert regularized offline RL framework.

Abstract

End-to-end (E2E) autonomous driving models that take only camera images as input and directly predict a future trajectory are appealing for their computational efficiency and potential for improved generalization via unified optimization; however, persistent failure modes remain due to reliance on imitation learning (IL). While online reinforcement learning (RL) could mitigate IL-induced issues, the computational burden of neural rendering-based simulation and large E2E networks renders iterative reward and hyperparameter tuning costly. We introduce a camera-only E2E offline RL framework that performs no additional exploration and trains solely on a fixed simulator dataset. Offline RL offers strong data efficiency and rapid experimental iteration, yet is susceptible to instability from overestimation on out-of-distribution (OOD) actions. To address this, we construct pseudo ground-truth trajectories from expert driving logs and use them as a behavior regularization signal, suppressing imitation of unsafe or suboptimal behavior while stabilizing value learning. Training and closed-loop evaluation are conducted in a neural rendering environment learned from the public nuScenes dataset. Empirically, the proposed method achieves substantial improvements in collision rate and route completion compared with IL baselines. Our code is available at https://github.com/ToyotaInfoTech/PEBC.