LLM-Guided Agentic Floor Plan Parsing for Accessible Indoor Navigation of Blind and Low-Vision People

arXiv cs.AI / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces an agentic LLM-guided framework that turns a single indoor floor plan image into a structured, retrievable knowledge base to support safe navigation for blind and low-vision (BLV) users without costly per-building infrastructure.
  • The approach uses a multi-agent parsing phase that builds a spatial knowledge graph via a self-correcting pipeline with iterative retry loops and corrective feedback.
  • A separate path-planning phase generates accessible navigation instructions while a Safety Evaluator agent checks for hazards along each proposed route.
  • Experiments on the UMBC Math and Psychology building (MP-1, MP-3) and the CVC-FP benchmark show higher success rates than the strongest single-call LLM baseline (Claude 3.7 Sonnet), especially for short and medium routes.
  • Overall, the results suggest the workflow improves reliability and scalability for accessible indoor navigation by combining structured parsing, planning, and safety evaluation.

Abstract

Indoor navigation remains a critical accessibility challenge for the blind and low-vision (BLV) individuals, as existing solutions rely on costly per-building infrastructure. We present an agentic framework that converts a single floor plan image into a structured, retrievable knowledge base to generate safe, accessible navigation instructions with lightweight infrastructure. The system has two phases: a multi-agent module that parses the floor plan into a spatial knowledge graph through a self-correcting pipeline with iterative retry loops and corrective feedback; and a Path Planner that generates accessible navigation instructions, with a Safety Evaluator agent assessing potential hazards along each route. We evaluate the system on the real-world UMBC Math and Psychology building (floors MP-1 and MP-3) and on the CVC-FP benchmark. On MP-1, we achieve success rates of 92.31%, 76.92%, and 61.54% for short, medium, and long routes, outperforming the strongest single-call baseline (Claude 3.7 Sonnet) at 84.62%, 69.23%, and 53.85%. On MP-3, we reach 76.92%, 61.54%, and 38.46%, compared to the best baseline at 61.54%, 46.15%, and 23.08%. These results show consistent gains over single-call LLM baselines and demonstrate that our workflow is a scalable solution for accessible indoor navigation for BLV individuals.

LLM-Guided Agentic Floor Plan Parsing for Accessible Indoor Navigation of Blind and Low-Vision People | AI Navigate