FineCog-Nav: Integrating Fine-grained Cognitive Modules for Zero-shot Multimodal UAV Navigation

arXiv cs.CV / 4/20/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • FineCog-Nav is a new top-down framework for UAV vision-language navigation that decomposes the task into fine-grained cognitive modules (language, perception, attention, memory, imagination, reasoning, and decision-making).
  • Each module uses a moderate-sized foundation model with role-specific prompts and well-defined structured protocols to improve coordination among modules and interpretability.
  • The work introduces AerialVLN-Fine, a new benchmark with 300 curated trajectories, sentence-level alignment between instructions and trajectories, and refined instructions that include explicit visual endpoints and landmark references.
  • Experiments report that FineCog-Nav improves zero-shot performance, particularly in instruction adherence, long-horizon planning, and generalization to previously unseen environments.
  • Overall, the authors argue that fine-grained cognitive modularization is an effective way to overcome limitations of existing zero-shot multimodal UAV navigation methods that rely on generic prompting and loosely coupled components.

Abstract

UAV vision-language navigation (VLN) requires an agent to navigate complex 3D environments from an egocentric perspective while following ambiguous multi-step instructions over long horizons. Existing zero-shot methods remain limited, as they often rely on large base models, generic prompts, and loosely coordinated modules. In this work, we propose FineCog-Nav, a top-down framework inspired by human cognition that organizes navigation into fine-grained modules for language processing, perception, attention, memory, imagination, reasoning, and decision-making. Each module is driven by a moderate-sized foundation model with role-specific prompts and structured input-output protocols, enabling effective collaboration and improved interpretability. To support fine-grained evaluation, we construct AerialVLN-Fine, a curated benchmark of 300 trajectories derived from AerialVLN, with sentence-level instruction-trajectory alignment and refined instructions containing explicit visual endpoints and landmark references. Experiments show that FineCog-Nav consistently outperforms zero-shot baselines in instruction adherence, long-horizon planning, and generalization to unseen environments. These results suggest the effectiveness of fine-grained cognitive modularization for zero-shot aerial navigation. Project page: https://smartdianlab.github.io/projects-FineCogNav.