Benchmarking Interaction, Beyond Policy: a Reproducible Benchmark for Collaborative Instance Object Navigation

arXiv cs.AI / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces QAsk-Nav, described as the first reproducible benchmark for Collaborative Instance Object Navigation (CoIN) that separately evaluates embodied navigation performance and human-style question-asking interaction.
  • QAsk-Nav is designed for partial observability and uses egocentric vision plus interactive natural-language dialogue, enabling agents to ask questions to resolve ambiguity between visually similar object instances.
  • The benchmark includes a lightweight, independently scored question-asking protocol, an enhanced navigation protocol with realistic diverse high-quality target descriptions, and an open-source dataset containing 28,000 quality-checked reasoning/question-asking traces.
  • Using QAsk-Nav, the authors present Light-CoNav, a lightweight unified model that is reported to be 3× smaller and 70× faster than prior modular approaches while achieving stronger generalization to unseen objects and environments.

Abstract

We propose Question-Asking Navigation (QAsk-Nav), the first reproducible benchmark for Collaborative Instance Object Navigation (CoIN) that enables an explicit, separate assessment of embodied navigation and collaborative question asking. CoIN tasks an embodied agent with reaching a target specified in free-form natural language under partial observability, using only egocentric visual observations and interactive natural-language dialogue with a human, where the dialogue can help to resolve ambiguity among visually similar object instances. Existing CoIN benchmarks are primarily focused on navigation success and offer no support for consistent evaluation of collaborative interaction. To address this limitation, QAsk-Nav provides (i) a lightweight question-asking protocol scored independently of navigation, (ii) an enhanced navigation protocol with realistic, diverse, high-quality target descriptions, and (iii) an open-source dataset, that includes 28,000 quality-checked reasoning and question-asking traces for training and analysis of interactive capabilities of CoIN models. Using the proposed QAsk-Nav benchmark, we develop Light-CoNav, a lightweight unified model for collaborative navigation that is 3x smaller and 70x faster than existing modular methods, while outperforming state-of-the-art CoIN approaches in generalization to unseen objects and environments. Project page at https://benchmarking-interaction.github.io/