Representative Spectral Correlation Network for Multi-source Remote Sensing Image Classification

arXiv cs.CV / 5/1/2026

💬 OpinionDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces RSCNet, a multi-source remote sensing image classification framework that fuses hyperspectral imagery with SAR/LiDAR data while addressing spectral redundancy and cross-source heterogeneity.
  • It proposes a Key Band Selection Module (KBSM) that adaptively selects task-relevant HSI bands using guidance from other sources, reducing redundancy and avoiding the information loss common in PCA-style spectral reduction.
  • It adds a Cross-source Adaptive Fusion Module (CAFM) that uses cross-source attention weighting plus local-global contextual refinement to improve feature interaction between modalities.
  • Experiments on three public benchmarks show RSCNet outperforms existing state-of-the-art methods while keeping computational complexity substantially lower, and the authors provide public code.

Abstract

Hyperspectral image (HSI) and SAR/LiDAR data offer complementary spectral and structural information for land-cover classification. However, their effective fusion remains challenging due to two major limitations: The spectral redundancy in high-dimensional HSI and the heterogeneous characteristics between multi-source data. To this end, we propose Representative Spectral Correlation Network (RSCNet), a novel multi-source image classification framework specifically designed to address the above challenges through spectral selection and adaptive interaction. The network incorporates two key components: (1) Key Band Selection Module (KBSM) that adaptively selects task-relevant spectral bands from the original HSI under cross-source guidance, thereby alleviating redundancy and mitigating information loss from conventional PCA-based spectral reduction. Moreover, the learned band subset exhibits highly discriminative spectral structures that align with discriminative semantic cues, promoting compact yet expressive representations. (2) Cross-source Adaptive Fusion Module (CAFM) that performs cross-source attention weighting and local-global contextual refinement to enhance cross-source feature interaction. Experiments on three public benchmark datasets demonstrate that our RSCNet achieves superior performance compared with state-of-the-art methods, while maintaining substantially lower computational complexity. Our codes are publicly available at https://github.com/oucailab/RSCNet.