ICR-Drive: Instruction Counterfactual Robustness for End-to-End Language-Driven Autonomous Driving

arXiv cs.CL / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ICR-Drive, a diagnostic framework to measure instruction counterfactual robustness for end-to-end language-conditioned autonomous driving agents.
  • It evaluates how language perturbations affect performance by generating instruction variants across four families: Paraphrase, Ambiguity, Noise, and Misleading (including goal-conflicting, authority-framed text).
  • Using controlled replay in CARLA with matched simulator settings and seeds, the framework isolates performance changes caused specifically by instruction wording rather than environmental randomness.
  • Results on LMDrive and BEVDriver indicate that even small instruction changes can cause large performance drops and distinct failure modes, highlighting a reliability gap for embodied foundation models in safety-critical contexts.

Abstract

Recent progress in vision-language-action (VLA) models has enabled language-conditioned driving agents to execute natural-language navigation commands in closed-loop simulation, yet standard evaluations largely assume instructions are precise and well-formed. In deployment, instructions vary in phrasing and specificity, may omit critical qualifiers, and can occasionally include misleading, authority-framed text, leaving instruction-level robustness under-measured. We introduce ICR-Drive, a diagnostic framework for instruction counterfactual robustness in end-to-end language-conditioned autonomous driving. ICR-Drive generates controlled instruction variants spanning four perturbation families: Paraphrase, Ambiguity, Noise, and Misleading, where Misleading variants conflict with the navigation goal and attempt to override intent. We replay identical CARLA routes under matched simulator configurations and seeds to isolate performance changes attributable to instruction language. Robustness is quantified using standard CARLA Leaderboard metrics and per-family performance degradation relative to the baseline instruction. Experiments on LMDrive and BEVDriver show that minor instruction changes can induce substantial performance drops and distinct failure modes, revealing a reliability gap for deploying embodied foundation models in safety-critical driving.