LLM4AD: Large Language Models for Autonomous Driving -- Concept, Review, Benchmark, Experiments, and Future Trends

arXiv cs.CL / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a new concept, LLM4AD, for using large language models to improve autonomous driving across perception, scene understanding, and interactive decision-making.
  • It reviews prior work on LLM4AD, then introduces a comprehensive evaluation benchmark spanning instruction-following and reasoning through LaMPilot-Bench, CARLA Leaderboard 1.0 (simulation), and NuPlanQA (multi-view VQA).
  • Extensive experiments are conducted on real autonomous vehicle platforms, comparing on-cloud versus on-edge LLM deployment for personalized decision-making and motion control.
  • The authors outline future directions that integrate vision-language diffusion models into driving, including a proposed ViLaD (Vision-Language Diffusion) framework.
  • Key remaining challenges are discussed, including latency, deployment constraints, security/privacy, safety, trust/transparency, and personalization.

Abstract

With the broader adoption and highly successful development of Large Language Models (LLMs), there has been growing interest and demand for applying LLMs to autonomous driving technology. Driven by their natural language understanding and reasoning capabilities, LLMs have the potential to enhance various aspects of autonomous driving systems, from perception and scene understanding to interactive decision-making. This paper first introduces the novel concept of designing Large Language Models for Autonomous Driving (LLM4AD), followed by a review of existing LLM4AD studies. Then, a comprehensive benchmark is proposed for evaluating the instruction-following and reasoning abilities of LLM4AD systems, which includes LaMPilot-Bench, CARLA Leaderboard 1.0 Benchmark in simulation and NuPlanQA for multi-view visual question answering. Furthermore, extensive real-world experiments are conducted on autonomous vehicle platforms, examining both on-cloud and on-edge LLM deployment for personalized decision-making and motion control. Next, the future trends of integrating language diffusion models into autonomous driving are explored, exemplified by the proposed ViLaD (Vision-Language Diffusion) framework. Finally, the main challenges of LLM4AD are discussed, including latency, deployment, security and privacy, safety, trust and transparency, and personalization.