Dual-Stage LLM Framework for Scenario-Centric Semantic Interpretation in Driving Assistance

arXiv cs.AI / 3/31/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a dual-stage, scenario-centric framework to audit LLM-based risk reasoning for driving assistance under reproducible, temporally bounded scenario windows derived from multimodal driving data.
  • It evaluates multiple models (two text-only and one multimodal) with fixed prompt constraints and a closed numeric risk schema to produce structured, comparable outputs.
  • Experiments show systematic inter-model divergence in how severity, evidence, and causal attribution are assigned, including different interpretations of vulnerable road user presence.
  • The authors argue that such variability can stem from intrinsic semantic ambiguity in risk interpretation—rather than isolated model malfunction—making ambiguity management a key requirement for safety-aligned ADAS deployments.

Abstract

Advanced Driver Assistance Systems (ADAS) increasingly rely on learning-based perception, yet safety-relevant failures often arise without component malfunction, driven instead by partial observability and semantic ambiguity in how risk is interpreted and communicated. This paper presents a scenario-centric framework for reproducible auditing of LLM-based risk reasoning in urban driving contexts. Deterministic, temporally bounded scenario windows are constructed from multimodal driving data and evaluated under fixed prompt constraints and a closed numeric risk schema, ensuring structured and comparable outputs across models. Experiments on a curated near-people scenario set compare two text-only models and one multimodal model under identical inputs and prompts. Results reveal systematic inter-model divergence in severity assignment, high-risk escalation, evidence use, and causal attribution. Disagreement extends to the interpretation of vulnerable road user presence, indicating that variability often reflects intrinsic semantic indeterminacy rather than isolated model failure. These findings highlight the importance of scenario-centric auditing and explicit ambiguity management when integrating LLM-based reasoning into safety-aligned driver assistance systems.