Cross-Stage Attention Propagation for Efficient Semantic Segmentation
arXiv cs.CV / 4/8/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that many lightweight semantic segmentation models use multi-scale decoders that recompute attention independently per feature scale, causing redundant computation because attention patterns across scales are highly correlated.
- It introduces Cross-Stage Attention Propagation (CSAP), which computes attention only at the deepest feature scale and then propagates those attention maps to shallower decoder stages, eliminating query-key attention computation there.
- CSAP is designed to maintain multi-scale contextual reasoning while substantially reducing decoder compute requirements.
- Reported results show strong efficiency and accuracy: CSAP-Tiny reaches 42.9% mIoU on ADE20K with 5.5 GFLOPs, 80.5% on Cityscapes with 21.5 GFLOPs, and 40.9% on COCO-Stuff 164K with 5.5 GFLOPs.
- The method outperforms SegNeXt-Tiny by +1.8% mIoU on ADE20K while using 16.8% fewer floating-point operations, suggesting a favorable accuracy/compute tradeoff for compact segmentation models.
Related Articles

Black Hat Asia
AI Business

The enforcement gap: why finding issues was never the problem
Dev.to

How I Built AI-Powered Auto-Redaction Into a Desktop Screenshot Tool
Dev.to

Agentic AI vs Traditional Automation: Why They Require Different Approaches in Modern Enterprises
Dev.to

Agentic AI vs Traditional Automation: Why Modern Enterprises Must Treat Them Differently
Dev.to