CodeGraphVLP: Code-as-Planner Meets Semantic-Graph State for Non-Markovian Vision-Language-Action Models

arXiv cs.RO / 4/27/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • CodeGraphVLP targets vision-language-action (VLA) robotics problems where long-horizon, non-Markovian environments require remembering earlier evidence that may later be occluded or hidden.
  • The framework combines a persistent semantic-graph state (tracking relevant entities and relations under partial observability) with an executable code-based hierarchical planner for subtask generation and progress checking.
  • It uses the planner’s subtask instructions and identified objects to create clutter-suppressed observations, improving visual grounding and reducing distraction for the VLA executor.
  • Experiments on real-world non-Markovian tasks show higher task completion than strong VLA baselines and history-enabled variants, while also reducing planning latency compared with VLM-in-the-loop approaches.
  • Extensive ablation studies validate the specific contribution of each component in the hierarchical semantic-graph + code-planner + progress-guided prompting pipeline.

Abstract

Vision-Language-Action (VLA) models promise generalist robot manipulation, but are typically trained and deployed as short-horizon policies that assume the latest observation is sufficient for action reasoning. This assumption breaks in non-Markovian long-horizon tasks, where task-relevant evidence can be occluded or appear only earlier in the trajectory, and where clutter and distractors make fine-grained visual grounding brittle. We present CodeGraphVLP, a hierarchical framework that enables reliable long-horizon manipulation by combining a persistent semantic-graph state with an executable code-based planner and progress-guided visual-language prompting. The semantic-graph maintains task-relevant entities and relations under partial observability. The synthesized planner executes over this semantic-graph to perform efficient progress checks and outputs a subtask instruction together with subtask-relevant objects. We use these outputs to construct clutter-suppressed observations that focus the VLA executor on critical evidence. On real-world non-Markovian tasks, CodeGraphVLP improves task completion over strong VLA baselines and history-enabled variants while substantially lowering planning latency compared to VLM-in-the-loop planning. We also conduct extensive ablation studies to confirm the contributions of each component.