Skill-Conditioned Visual Geolocation for Vision-Language

arXiv cs.CV / 4/13/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces GeoSkill, a training-free vision-language geolocation framework that adds more structured geographic reasoning than existing approaches relying on implicit parametric memory.
  • GeoSkill initializes a Skill-Graph by converting human expert geolocation trajectories into atomic, natural-language skills, enabling inference to be guided by explicit skill representations.
  • An Autonomous Evolution mechanism uses a larger model to run multiple reasoning rollouts on web-derived image-coordinate pairs, then synthesizes and prunes skills based on both successful and failed trajectories to reduce bias.
  • Experiments on GeoRC show GeoSkill improves both geolocation accuracy and reasoning faithfulness while maintaining strong generalization to external datasets.
  • The approach claims to enable self-evolution and the emergence of novel, verifiable skills without any parameter updates, aiming to better capture real-world geographic knowledge.

Abstract

Vision-language models (VLMs) have shown a promising ability in image geolocation, but they still lack structured geographic reasoning and the capacity for autonomous self-evolution. Existing methods predominantly rely on implicit parametric memory, which often exploits outdated knowledge and generates hallucinated reasoning. Furthermore, current inference is a "one-off" process, lacking the feedback loops necessary for self-evolution based on reasoning outcomes. To address these issues, we propose GeoSkill, a training-free framework based on an evolving Skill-Graph. We first initialize the graph by refining human expert trajectories into atomic, natural-language skills. For execution, GeoSkill employs an inference model to perform direct reasoning guided by the current Skill-Graph. For continuous growth, an Autonomous Evolution mechanism leverages a larger model to conduct multiple reasoning rollouts on image-coordinate pairs sourced from web-scale data and verified real-world reasoning. By analyzing both successful and failed trajectories from these rollouts, the mechanism iteratively synthesizes and prunes skills, effectively expanding the Skill-Graph and correcting geographic biases without any parameter updates. Experiments demonstrate that GeoSkill achieves promising performance in both geolocation accuracy and reasoning faithfulness on GeoRC, while maintaining superior generalization across diverse external datasets. Furthermore, our autonomous evolution fosters the emergence of novel, verifiable skills, significantly enhancing the system's cognition of real-world geographic knowledge beyond isolated case studies.