ProVG: Progressive Visual Grounding via Language Decoupling for Remote Sensing Imagery

arXiv cs.CV / 4/3/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • ProVG (Progressive Visual Grounding) targets remote sensing visual grounding by improving sentence-level alignment with fine-grained linguistic cues like spatial relations and object attributes.
  • The method decouples language into global context, spatial relations, and object attributes, then integrates them using a progressive cross-modal modulator with a survey–locate–verify (coarse-to-fine) attention scheme.
  • To handle remote-sensing-specific challenges, ProVG adds cross-scale fusion for large-scale variability and a language-guided calibration decoder to refine alignment during prediction.
  • It uses a unified multi-task head to support both referring expression comprehension and segmentation, and reports state-of-the-art results on RRSIS-D and RISBench.
  • The work introduces a stage-aware way to use different linguistic components across the grounding pipeline, yielding consistent performance gains over prior approaches.

Abstract

Remote sensing visual grounding (RSVG) aims to localize objects in remote sensing imagery according to natural language expressions. Previous methods typically rely on sentence-level vision-language alignment, which struggles to exploit fine-grained linguistic cues, such as \textit{spatial relations} and \textit{object attributes}, that are crucial for distinguishing objects with similar characteristics. Importantly, these cues play distinct roles across different grounding stages and should be leveraged accordingly to provide more explicit guidance. In this work, we propose \textbf{ProVG}, a novel RSVG framework that improves localization accuracy by decoupling language expressions into global context, spatial relations, and object attributes. To integrate these linguistic cues, ProVG employs a simple yet effective progressive cross-modal modulator, which dynamically modulates visual attention through a \textit{survey-locate-verify} scheme, enabling coarse-to-fine vision-language alignment. In addition, ProVG incorporates a cross-scale fusion module to mitigate the large-scale variations in remote sensing imagery, along with a language-guided calibration decoder to refine cross-modal alignment during prediction. A unified multi-task head further enables ProVG to support both referring expression comprehension and segmentation tasks. Extensive experiments on two benchmarks, \textit{i.e.}, RRSIS-D and RISBench, demonstrate that ProVG consistently outperforms existing methods, achieving new state-of-the-art performance.