StateVLM: A State-Aware Vision-Language Model for Robotic Affordance Reasoning

arXiv cs.CV / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • StateVLM is introduced as a state-aware vision-language model for robotic affordance reasoning, targeting VLM weaknesses in numerical reasoning such as object detection and state localization.
  • The paper proposes a fine-tuning strategy that uses box decoder outputs to compute an Auxiliary Regression Loss (ARL) while keeping standard sequence prediction at inference.
  • By framing numerical reasoning as a regression task, the approach aims to learn fine-grained object representations including precise localization, object states, and graspable regions.
  • The authors create an open-source benchmark called OSAR (Object State Affordance Reasoning) with 1,172 scenes, 7,746 objects, and corresponding bounding boxes to evaluate object-state reasoning.
  • Experiments show that adding ARL yields an average performance improvement of 1.6% on adapted benchmarks and 5.2% on OSAR, with ARL also improving output consistency on complex affordance reasoning.

Abstract

Vision-language models (VLMs) have shown remarkable performance in various robotic tasks, as they can perceive visual information and understand natural language instructions. However, when applied to robotics, VLMs remain subject to a fundamental limitation inherent in large language models (LLMs): they struggle with numerical reasoning, particularly in object detection and object-state localization. To explore numerical reasoning as a regression task in VLMs, we propose a novel training strategy to adapt VLMs for object detection and object-state localization. This approach leverages box decoder outputs to compute an Auxiliary Regression Loss (ARL) during fine-tuning, while preserving standard sequence prediction at inference. We leverage this training strategy to develop StateVLM (State-aware Vision-Language Model), a novel model designed to perceive and learn fine-grained object representations, including precise localization of objects and their states, as well as graspable regions. Due to the lack of a benchmark for object-state affordance reasoning, we introduce an open-source benchmark, Object State Affordance Reasoning (OSAR), which contains 1,172 scenes with 7,746 individual objects and corresponding bounding boxes. Comparative experiments on adapted benchmarks (RefCOCO, RefCOCO+, and \mbox{RefCOCOg}) demonstrate that ARL improves model performance by an average of 1.6\% compared to models without ARL. Experiments on the OSAR benchmark further support this finding, showing that StateVLM with ARL achieves an average of 5.2\% higher performance than models without ARL. In particular, ARL is also important for the complex task of affordance reasoning in OSAR, where it enhances the consistency of model outputs.