STAR: Mitigating Cascading Errors in Spatial Reasoning via Turn-point Alignment and Segment-level DPO

arXiv cs.CV / 4/2/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces STAR, a two-stage framework for mitigating cascading errors in LLM-based structured spatial navigation by using topological anchors and turn-point alignment.
  • STAR’s first stage applies supervised fine-tuning to internalize spatial semantics and prune redundant paths that commonly lead to early mistakes compounding later.
  • Its second stage uses Spatial-aware Segment-level Direct Preference Optimization (SDPO) to improve self-correction during long-horizon navigation.
  • The authors release the RedMaze-23K dataset, featuring human-inspired turn-point annotations designed to better support spatial reasoning training and evaluation.
  • Experiments report state-of-the-art results among open-source models, with a 32B STAR variant surpassing DeepSeek-V3 (29.27% vs. 25.00%) and achieving 82.4% of GPT-4 performance.

Abstract

Structured spatial navigation is a core benchmark for Large Language Models (LLMs) spatial reasoning. Existing paradigms like Visualization-of-Thought (VoT) are prone to cascading errors in complex topologies. To solve this, we propose STAR, a two-stage framework grounded on topological anchors, and introduce the RedMaze-23K dataset with human-inspired turnpoint annotations. The first stage uses supervised fine-tuning to help models internalize spatial semantics and prune redundant paths. The second adopts Spatial-aware Segment-level Direct Preference Optimization (SDPO) to refine self-correction in long-horizon navigation. Experiments show STAR achieves state-of-the-art performance among open-source models: its 32B variant outperforms DeepSeek-V3 (29.27% vs. 25.00%) and reaches 82.4% of GPT-4's performance.