Enhanced Swin Transformer and Edge Spatial attention For Remote Sensing Image Semantic Segmentation

Fuxiang Liu, Zhiqiang Hu, Lei Li*, Hanlu Li, Xinxin Liu

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Combining convolutional neural networks (CNNs) and transformers is a crucial direction in remote sensing image semantic segmentation. However, due to differences in the spatial information focus and feature extraction methods, existing feature transfer and fusion strategies do not effectively integrate the advantages of both approaches. To address these issues, we propose a CNN-transformer hybrid network for precise remote sensing image semantic segmentation. We propose a novel Swin Transformer block to optimize feature extraction and enable the model to handle remote sensing images of arbitrary sizes. Additionally, we design an Edge Spatial Attention module to focus attention on local edge structures, effectively integrating global features and local details. This facilitates efficient information flow between the Transformer encoder and CNN decoder. Finally, a multi-scale convolutional decoder is employed to fully leverage both global information from the Transformer and local features from the CNN, leading to accurate segmentation results. Our network achieved state-of-the-art performance on the Vaihingen and Potsdam datasets, reaching mIoU and F1 scores of 67.37% and 79.82%, as well as 72.39% and 83.68%, respectively. Our code is publicly available at: http://github.com/TarsDolores/LZ.

Original languageEnglish
JournalIEEE Signal Processing Letters
DOIs
Publication statusAccepted/In press - 2025

Keywords

  • edge detection
  • remote sensing image
  • semantic segmentation
  • swin transformer

Fingerprint

Dive into the research topics of 'Enhanced Swin Transformer and Edge Spatial attention For Remote Sensing Image Semantic Segmentation'. Together they form a unique fingerprint.

Cite this