DB SwinT: A Dual-Branch Swin Transformer Network for Road Extraction in Optical Remote Sensing Imagery

arXiv cs.CV / 3/26/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DB SwinT, a dual-branch Swin Transformer network designed to improve road extraction from optical remote sensing imagery under complex occlusions.
  • It combines Swin Transformer long-range dependency modeling with a U-Net-style multi-scale feature fusion pipeline to better recover both fine road structures and overall network continuity.
  • A dual-branch encoder learns complementary local (fine details in occluded regions) and global (broader semantic context) representations, addressing fragmented road outputs.
  • An Attentional Feature Fusion (AFF) module adaptively fuses the two branches to enhance detection of occluded road segments.
  • Experiments on Massachusetts and DeepGlobe report IoU scores of 79.35% and 74.84%, respectively, indicating improved performance for remote sensing road extraction.

Abstract

With the continuous improvement in the spatial resolution of optical remote sensing imagery, accurate road extraction has become increasingly important for applications such as urban planning, traffic monitoring, and disaster management. However, road extraction in complex urban and rural environments remains challenging, as roads are often occluded by trees, buildings, and other objects, leading to fragmented structures and reduced extraction accuracy. To address this problem, this paper proposes a Dual-Branch Swin Transformer network (DB SwinT) for road extraction. The proposed framework combines the long-range dependency modeling capability of the Swin Transformer with the multi-scale feature fusion strategy of U-Net, and employs a dual-branch encoder to learn complementary local and global representations. Specifically, the local branch focuses on recovering fine structural details in occluded areas, while the global branch captures broader semantic context to preserve the overall continuity of road networks. In addition, an Attentional Feature Fusion (AFF) module is introduced to adaptively fuse features from the two branches, further enhancing the representation of occluded road segments. Experimental results on the Massachusetts and DeepGlobe datasets show that DB SwinT achieves Intersection over Union (IoU) scores of 79.35\% and 74.84\%, respectively, demonstrating its effectiveness for road extraction from optical remote sensing imagery.

DB SwinT: A Dual-Branch Swin Transformer Network for Road Extraction in Optical Remote Sensing Imagery | AI Navigate