Cosine-Normalized Attention for Hyperspectral Image Classification

arXiv cs.CV / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that standard dot-product attention in transformer-based hyperspectral image classification can be suboptimal because it entangles feature magnitude with direction (orientation) rather than capturing the angular structure of hyperspectral signatures.
  • It proposes cosine-normalized attention by projecting query and key embeddings onto a unit hypersphere and using squared cosine similarity to emphasize angular relationships while reducing sensitivity to magnitude changes.
  • The method is incorporated into a spatial-spectral Transformer and tested in an extremely limited-supervision setting.
  • Experiments on three benchmark datasets show consistent performance gains, outperforming multiple recent Transformer- and Mamba-based approaches even with a lightweight backbone.
  • Ablation/controlled analyses comparing different attention score functions indicate that cosine-based scoring provides a beneficial inductive bias for hyperspectral representation learning.

Abstract

Transformer-based methods have improved hyperspectral image classification (HSIC) by modeling long-range spatial-spectral dependencies; however, their attention mechanisms typically rely on dot-product similarity, which mixes feature magnitude and orientation and may be suboptimal for hyperspectral data. This work revisits attention scoring from a geometric perspective and introduces a cosine-normalized attention formulation that aligns similarity computation with the angular structure of hyperspectral signatures. By projecting query and key embeddings onto a unit hypersphere and applying a squared cosine similarity, the proposed method emphasizes angular relationships while reducing sensitivity to magnitude variations. The formulation is integrated into a spatial-spectral Transformer and evaluated under extremely limited supervision. Experiments on three benchmark datasets demonstrate that the proposed approach consistently achieves higher performance, outperforming several recent Transformer- and Mamba-based models despite using a lightweight backbone. In addition, a controlled analysis of multiple attention score functions shows that cosine-based scoring provides a reliable inductive bias for hyperspectral representation learning.