Cross-Stage Attention Propagation for Efficient Semantic Segmentation

arXiv cs.CV / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that many lightweight semantic segmentation models use multi-scale decoders that recompute attention independently per feature scale, causing redundant computation because attention patterns across scales are highly correlated.
  • It introduces Cross-Stage Attention Propagation (CSAP), which computes attention only at the deepest feature scale and then propagates those attention maps to shallower decoder stages, eliminating query-key attention computation there.
  • CSAP is designed to maintain multi-scale contextual reasoning while substantially reducing decoder compute requirements.
  • Reported results show strong efficiency and accuracy: CSAP-Tiny reaches 42.9% mIoU on ADE20K with 5.5 GFLOPs, 80.5% on Cityscapes with 21.5 GFLOPs, and 40.9% on COCO-Stuff 164K with 5.5 GFLOPs.
  • The method outperforms SegNeXt-Tiny by +1.8% mIoU on ADE20K while using 16.8% fewer floating-point operations, suggesting a favorable accuracy/compute tradeoff for compact segmentation models.

Abstract

Recent lightweight semantic segmentation methods have made significant progress by combining compact backbones with efficient decoder heads. However, most multi-scale decoders compute attention independently at each feature scale, introducing substantial redundancy since the resulting attention distributions across scales are strongly correlated. We propose Cross-Stage Attention Propagation (CSAP), a decoder framework that computes attention at the deepest feature scale and propagates the resulting attention maps to shallower stages, bypassing query-key computation at those stages entirely. This design preserves multi-scale contextual reasoning while substantially reducing the decoder's computational cost. CSAP-Tiny achieves 42.9% mIoU on ADE20K with only 5.5 GFLOPs, 80.5% on Cityscapes with 21.5 GFLOPs, and 40.9% on COCO-Stuff 164K with 5.5 GFLOPs, surpassing SegNeXt-Tiny by +1.8% on ADE20K while requiring 16.8% fewer floating-point operations.