Think before Go: Hierarchical Reasoning for Image-goal Navigation

arXiv cs.RO / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses image-goal navigation, where an agent must reach a target location specified by an image in unseen environments, noting that end-to-end policies struggle when the goal is far away or in another room.
  • It introduces Hierarchical Reasoning Navigation (HRNav), which splits the problem into high-level planning and low-level execution to better handle long-horizon navigation.
  • For high-level planning, HRNav trains a vision-language model on a self-collected dataset to produce short-horizon instructions (e.g., whether to go through a door or proceed down a hallway).
  • For low-level execution, it uses an online reinforcement learning policy that selects actions based on the short-horizon plan, and it adds a Wandering Suppression Penalty (WSP) to reduce aimless wandering.
  • Experiments in simulation and real-world settings show that HRNav outperforms existing approaches, validating the hierarchical design and wandering mitigation strategy.

Abstract

Image-goal navigation steers an agent to a target location specified by an image in unseen environments. Existing methods primarily handle this task by learning an end-to-end navigation policy, which compares the similarities of target and observation images and directly predicts the actions. However, when the target is distant or lies in another room, such methods fail to extract informative visual cues, leading the agent to wander around. Motivated by the human cognitive principle that deliberate, high-level reasoning guides fast, reactive execution in complex tasks, we propose Hierarchical Reasoning Navigation (HRNav), a framework that decomposes image-goal navigation into high-level planning and low-level execution. In high-level planning, a vision-language model is trained on a self-collected dataset to generate a short-horizon plan, such as whether the agent should walk through the door or down the hallway. This downgrades the difficulty of the long-horizon task, making it more amenable to the execution part. In low-level execution, an online reinforcement learning policy is utilized to decide actions conditioned on the short-horizon plan. We also devise a novel Wandering Suppression Penalty (WSP) to further reduce the wandering problem. Together, these components form a hierarchical framework for Image-Goal Navigation. Extensive experiments in both simulation and real-world environments demonstrate the superiority of our method.