From Prompt to Physical Actuation: Holistic Threat Modeling of LLM-Enabled Robotic Systems

arXiv cs.RO / 5/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper warns that when LLMs are used in autonomous robots for planning and control, malicious or unsafe prompts/outputs can propagate through the decision pipeline and cause real-world physical harm.
  • It proposes a unified architectural threat model for an edge-cloud LLM-enabled robot using a hierarchical Data Flow Diagram and STRIDE-per-interaction analysis.
  • By analyzing six boundary-crossing interaction points with a taxonomy covering conventional cyber threats, adversarial threats, and conversational threats, the study shows these threat types converge at the same boundaries.
  • The authors trace three cross-boundary attack chains that can ultimately lead to unsafe actuation, highlighting architectural weaknesses such as missing semantic validation, risky cross-modal translation (vision to language instructions), and insufficient mediation during provider-side tool use.
  • The work claims novelty as the first DFD-based approach that integrates all three threat categories across the full perception–planning–actuation pipeline for LLM-enabled robotic systems.

Abstract

As large language models are integrated into autonomous robotic systems for task planning and control, compromised inputs or unsafe model outputs can propagate through the planning pipeline to physical-world consequences. Although prior work has studied robotic cybersecurity, adversarial perception attacks, and LLM safety independently, no existing study traces how these threat categories interact and propagate across trust boundaries in a unified architectural model. We address this gap by modeling an LLM-enabled autonomous robot in an edge-cloud architecture as a hierarchical Data Flow Diagram and applying STRIDE-per-interaction analysis across six boundary-crossing interaction points using a three-category taxonomy of Conventional Cyber Threats, Adversarial Threats, and Conversational Threats. The analysis reveals that these categories converge at the same boundary crossings, and we trace three cross-boundary attack chains from external entry points to unsafe physical actuation, each exposing a distinct architectural property: the absence of independent semantic validation between user input and actuator dispatch, cross-modal translation from visual perception to language-model instruction, and unmediated boundary crossing through provider-side tool use. To our knowledge, this is the first DFD-based threat analysis integrating all three threat categories across the full perception-planning-actuation pipeline of an LLM-enabled robotic system.