Physics-Aligned Spectral Mamba: Decoupling Semantics and Dynamics for Few-Shot Hyperspectral Target Detection

arXiv cs.CV / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes SpecMamba, a parameter-efficient, frequency-aware meta-learning framework for few-shot hyperspectral target detection that targets overfitting and inefficiency in full-parameter fine-tuning.
  • It introduces a Discrete Cosine Transform Mamba Adapter (DCTMA) that projects spectral features into the frequency domain and uses Mamba’s linear-complexity state-space recursion to model global spectral dependencies and band continuity.
  • To reduce prototype drift under limited samples, the method adds a Prior-Guided Tri-Encoder (PGTE) that leverages laboratory spectral priors to guide adapter training while keeping frozen Transformer semantic representations stable.
  • For test-time adaptation, it develops Self-Supervised Pseudo-Label Mapping (SSPLM) using uncertainty-aware sampling and dual-path consistency constraints to refine decision boundaries efficiently.
  • Experiments on multiple public hyperspectral datasets show SpecMamba improves detection accuracy and cross-domain generalization over existing state-of-the-art approaches.

Abstract

Meta-learning facilitates few-shot hyperspectral target detection (HTD), but adapting deep backbones remains challenging. Full-parameter fine-tuning is inefficient and prone to overfitting, and existing methods largely ignore the frequency-domain structure and spectral band continuity of hyperspectral data, limiting spectral adaptation and cross-domain generalization.To address these challenges, we propose SpecMamba, a parameter-efficient and frequency-aware framework that decouples stable semantic representation from agile spectral adaptation. Specifically, we introduce a Discrete Cosine Transform Mamba Adapter (DCTMA) on top of frozen Transformer representations. By projecting spectral features into the frequency domain via DCT and leveraging Mamba's linear-complexity state-space recursion, DCTMA explicitly captures global spectral dependencies and band continuity while avoiding the redundancy of full fine-tuning. Furthermore, to address prototype drift caused by limited sample sizes, we design a Prior-Guided Tri-Encoder (PGTE) that allows laboratory spectral priors to guide the optimization of the learnable adapter without disrupting the stable semantic feature space. Finally, a Self-Supervised Pseudo-Label Mapping (SSPLM) strategy is developed for test-time adaptation, enabling efficient decision boundary refinement through uncertainty-aware sampling and dual-path consistency constraints. Extensive experiments on multiple public datasets demonstrate that SpecMamba consistently outperforms state-of-the-art methods in detection accuracy and cross-domain generalization.