ICR-Drive: Instruction Counterfactual Robustness for End-to-End Language-Driven Autonomous Driving
arXiv cs.CL / 4/8/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces ICR-Drive, a diagnostic framework to measure instruction counterfactual robustness for end-to-end language-conditioned autonomous driving agents.
- It evaluates how language perturbations affect performance by generating instruction variants across four families: Paraphrase, Ambiguity, Noise, and Misleading (including goal-conflicting, authority-framed text).
- Using controlled replay in CARLA with matched simulator settings and seeds, the framework isolates performance changes caused specifically by instruction wording rather than environmental randomness.
- Results on LMDrive and BEVDriver indicate that even small instruction changes can cause large performance drops and distinct failure modes, highlighting a reliability gap for embodied foundation models in safety-critical contexts.
Related Articles

Meta's latest model is as open as Zuckerberg's private school
The Register

Why multi-agent AI security is broken (and the identity patterns that actually work)
Dev.to
BANKING77-77: New best of 94.61% on the official test set (+0.13pp) over our previous tests 94.48%.
Reddit r/artificial
A Comprehensive Implementation Guide to ModelScope for Model Search, Inference, Fine-Tuning, Evaluation, and Export
MarkTechPost

Harness Engineering: The Next Evolution of AI Engineering
Dev.to