A Unified Foundation Model for All-in-One Multi-Modal Remote Sensing Image Restoration and Fusion with Language Prompting

arXiv cs.CV / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces LLaRS, a “unified foundation model” for multi-modal remote sensing image restoration and fusion that uses language prompting to handle multiple low-level vision tasks in one framework.
  • It addresses sensor heterogeneity by applying Sinkhorn-Knopp optimal transport to align heterogeneous bands into semantically matched slots before processing.
  • LLaRS uses three mixture-of-experts components—convolutional experts for spatial patterns, channel-mixing experts for spectral fidelity, and attention experts with low-rank adapters for global context—to improve performance across degradation types.
  • Training relies on a new million-scale multi-task dataset (LLaRS1M) covering eleven tasks using both real paired observations and controlled synthetic degradations, with diverse natural-language prompts for conditioning.
  • Experiments report that LLaRS outperforms seven baselines consistently, and parameter-efficient fine-tuning shows strong transfer/adaptation on unseen data, with code provided via the project repository.

Abstract

Remote sensing imagery suffers from clouds, haze, noise, resolution limits, and sensor heterogeneity. Existing restoration and fusion approaches train separate models per degradation type. In this work, we present Language-conditioned Large-scale Remote Sensing restoration model (LLaRS), the first unified foundation model for multi-modal and multi-task remote sensing low-level vision. LLaRS employs Sinkhorn-Knopp optimal transport to align heterogeneous bands into semantically matched slots, routes features through three complementary mixture-of-experts layers (convolutional experts for spatial patterns, channel-mixing experts for spectral fidelity, and attention experts with low-rank adapters for global context), and stabilizes joint training via step-level dynamic weight adjustment. To train LLaRS, we construct LLaRS1M, a million-scale multi-task dataset spanning eleven restoration and enhancement tasks, integrating real paired observations and controlled synthetic degradations with diverse natural language prompts. Experiments show LLaRS consistently outperforms seven competitive models, and parameter-efficient finetuning experiments demonstrate strong transfer capability and adaptation efficiency on unseen data. Repo: https://github.com/yc-cui/LLaRS