Continual Vision-Language Learning for Remote Sensing: Benchmarking and Analysis

arXiv cs.CV / 4/2/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies a key limitation of remote sensing vision-language models: they are trained on static datasets and struggle to adapt to newly emerging sensing modalities and downstream tasks without catastrophic forgetting.
  • It introduces CLeaRS, a new benchmark for continual vision-language learning in remote sensing, containing 10 curated subsets and 207k+ image-text pairs across diverse tasks, modalities, and application scenarios.
  • The authors define three evaluation protocols—long-horizon, modality-incremental, and task-incremental—to measure how well models retain and adapt over time.
  • Extensive experiments show catastrophic forgetting occurs across all continual learning settings tested, and adapted continual learning methods provide limited gains for transitions involving tasks, instructions, and modalities.
  • The results motivate the development of continual learning approaches specifically tailored to the remote sensing vision-language setting.

Abstract

Current remote sensing vision-language models (RS VLMs) demonstrate impressive performance in image interpretation but rely on static training data, limiting their ability to accommodate continuously emerging sensing modalities and downstream tasks. This exposes a fundamental challenge: enabling RS VLMs to continually adapt without catastrophic forgetting. Despite its practical importance, the continual learning capability of RS VLMs remains underexplored, and no dedicated benchmark currently exists. In this work, we present CLeaRS, a comprehensive benchmark for continual vision-language learning in remote sensing. CLeaRS comprises 10 curated subsets with over 207k image-text pairs, spanning diverse interpretation tasks, sensing modalities, and application scenarios. We further define three evaluation protocols: long-horizon, modality-incremental, and task-incremental settings, to systematically assess continual adaptation. Extensive benchmarking of diverse vision-language models reveals catastrophic forgetting across all settings. Moreover, representative continual learning methods, when adapted to RS VLMs, exhibit limited effectiveness in handling task, instruction, and modality transitions. Our findings underscore the need for developing continual learning methods tailored to RS VLMs.