XAttnRes: Cross-Stage Attention Residuals for Medical Image Segmentation
arXiv cs.CV / 4/7/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces Cross-Stage Attention Residuals (XAttnRes), a new mechanism that keeps a global feature-history pool of prior encoder and decoder stage outputs for medical image segmentation.
- XAttnRes uses lightweight pseudo-query attention so each stage can selectively aggregate information from all preceding representations, improving over fixed residual connections.
- It adds spatial alignment and channel projection to bridge dimensional and resolution differences between LLM-style same-dimensional layers and multi-scale encoder-decoder segmentation architectures, with minimal added overhead.
- Experiments on four datasets across three imaging modalities show consistent segmentation gains when XAttnRes is incorporated into existing models.
- The authors report that XAttnRes can achieve baseline-competitive results even without traditional skip connections, implying learned attention-based aggregation can replace some inter-stage information flow.
