AnyImageNav: Any-View Geometry for Precise Last-Meter Image-Goal Navigation

arXiv cs.RO / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • AnyImageNav addresses Image Goal Navigation’s coarse stopping metric by enabling precise 6-DoF camera pose recovery needed for downstream manipulation tasks.
  • The method treats the goal image as a geometric query, registering it to agent observations via dense pixel-level correspondences to recover an accurate pose.
  • It uses a semantic-to-geometric cascade: a semantic relevance signal drives exploration and only triggers a 3D multi-view foundation model when views are highly relevant to the goal.
  • The foundation model then self-certifies registration iteratively to ensure accurate pose recovery rather than relying on adapted baselines.
  • Reported results set new state-of-the-art navigation success on Gibson (93.1%) and HM3D (82.6%) while improving pose error by 5–10x versus adapted baselines.

Abstract

Image Goal Navigation (ImageNav) is evaluated by a coarse success criterion, the agent must stop within 1m of the target, which is sufficient for finding objects but falls short for downstream tasks such as grasping that require precise positioning. We introduce AnyImageNav, a training-free system that pushes ImageNav toward this more demanding setting. Our key insight is that the goal image can be treated as a geometric query: any photo of an object, a hallway, or a room corner can be registered to the agent's observations via dense pixel-level correspondences, enabling recovery of the exact 6-DoF camera pose. Our method realizes this through a semantic-to-geometric cascade: a semantic relevance signal guides exploration and acts as a proximity gate, invoking a 3D multi-view foundation model only when the current view is highly relevant to the goal image; the model then self-certifies its registration in a loop for an accurate recovered pose. Our method sets state-of-the-art navigation success rates on Gibson (93.1%) and HM3D (82.6%), and achieves pose recovery that prior methods do not provide: a position error of 0.27m and heading error of 3.41 degrees on Gibson, and 0.21m / 1.23 degrees on HM3D, a 5-10x improvement over adapted baselines.