BridgeSim: Unveiling the OL-CL Gap in End-to-End Autonomous Driving
arXiv cs.RO / 4/14/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates the OL-CL gap in end-to-end autonomous driving, showing that open-loop (OL) policies that score well in OL evaluation can fail when deployed in closed-loop (CL) settings.
- It attributes the gap primarily to Observational Domain Shift (largely recoverable via adaptation) and Objective Mismatch (a more structural problem that limits modeling of complex reactive behaviors).
- The authors find many OL policies learn a biased Q-value estimator that overlooks CL reactivity and lacks the temporal awareness needed to prevent compounding errors.
- They propose a test-time adaptation (TTA) framework that calibrates observational shift, reduces state-action biases, and enforces temporal consistency.
- Experiments indicate TTA reduces planning biases and improves scaling dynamics, while also revealing that common OL evaluation protocols can miss closed-loop deployment “blind spots.”
Related Articles

Black Hat Asia
AI Business
Microsoft launches MAI-Image-2-Efficient, a cheaper and faster AI image model
VentureBeat

The AI School Bus Camera Company Blanketing America in Tickets
Dev.to
GPT-5.3 and GPT-5.4 on OpenClaw: Setup and Configuration...
Dev.to
GLM-5 on OpenClaw: Setup Guide, Benchmarks, and When to...
Dev.to