Region Matters: Efficient and Reliable Region-Aware Visual Place Recognition

arXiv cs.CV / 4/27/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces FoL++, a region-aware Visual Place Recognition (VPR) approach designed to reduce perceptual aliasing from irrelevant regions and improve inefficient re-ranking.
  • FoL++ adds a Reliability Estimation Branch that produces spatial reliability maps explicitly modeling occlusion resistance, helping weight local matches more effectively.
  • Two spatial alignment losses (SAL and SCEL) are used to align features and emphasize salient regions for more reliable region-level representations.
  • It uses weakly supervised training via a pseudo-correspondence strategy for dense local supervision, along with an Adaptive Candidate Scheduler that resizes candidate pools based on global similarity.
  • Experiments on seven benchmarks show state-of-the-art results with a lightweight memory footprint, including a reported 40% faster inference than FoL, and the code/model are planned for release and merging with FoL.

Abstract

Visual Place Recognition (VPR) determines a query image's geographic location by matching it against geotagged databases. However, existing methods struggle with perceptual aliasing caused by irrelevant regions and inefficient re-ranking due to rigid candidate scheduling. To address these issues, we introduce FoL++, a method combining robust discriminative region modeling with adaptive re-ranking. Specifically, we propose a Reliability Estimation Branch to generate spatial reliability maps that explicitly model occlusion resistance. This representation is further optimized by two spatial alignment losses (SAL and SCEL) to effectively align features and highlight salient regions. For weakly supervised learning without manual annotations, a pseudo-correspondence strategy generates dense local feature supervision directly from aggregation clusters. Our Adaptive Candidate Scheduler dynamically resizes candidate pools based on global similarity. By weighting local matches by reliability and adaptively fusing global and local evidence, FoL++ surpasses traditional independent matching systems. Extensive experiments across seven benchmarks demonstrate that FoL++ achieves state-of-the-art performance with a lightweight memory footprint, improving inference speed by 40% over FoL. Code and models will be released (and merged with FoL) at https://github.com/chenshunpeng/FoL.