VeriGraph: Scene Graphs for Execution Verifiable Robot Planning

arXiv cs.RO / 4/20/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces VeriGraph, a framework that uses vision-language models (VLMs) to improve robot task planning, which otherwise often yields incorrect action sequences.
  • VeriGraph converts input images into scene graphs as an intermediate representation, capturing key objects and spatial relationships to support more reliable verification.
  • The system iteratively checks and corrects action sequences produced by an LLM-based task planner to ensure feasibility and that constraints are respected.
  • Experiments across multiple manipulation scenarios show large improvements over baseline methods, including +58% on language-based tasks, +56% on tangram puzzle tasks, and +30% on image-based tasks.
  • The authors provide code and qualitative results at the project website for further inspection and reuse.

Abstract

Recent progress in vision-language models (VLMs) has opened new possibilities for robot task planning, but these models often produce incorrect action sequences. To address these limitations, we propose VeriGraph, a novel framework that integrates VLMs for robotic planning while verifying action feasibility. VeriGraph uses scene graphs as an intermediate representation to capture key objects and spatial relationships, enabling more reliable plan verification and refinement. The system generates a scene graph from input images and uses it to iteratively check and correct action sequences generated by an LLM-based task planner, ensuring constraints are respected and actions are executable. Our approach significantly enhances task completion rates across diverse manipulation scenarios, outperforming baseline methods by 58% on language-based tasks, 56% on tangram puzzle tasks, and 30% on image-based tasks. Qualitative results and code can be found at https://verigraph-agent.github.io.