2022
DOI: 10.1109/jstars.2022.3217038
|View full text |Cite
|
Sign up to set email alerts
|

STransUNet: A Siamese TransUNet-Based Remote Sensing Image Change Detection Network

Abstract: In modern remote sensing image change detection (CD), convolution Neural Network (CNN), especially U-shaped structure (UNet), has achieved great success due to powerful discriminative ability. However, UNet-based CNN networks usually have limitations in modeling global dependencies due to intrinsic locality of convolution operations. Transformer has recently emerged as an alternative architecture for dense prediction tasks due to global self-attention mechanism. However, due to limitation of hardware resources… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(7 citation statements)
references
References 45 publications
0
7
0
Order By: Relevance
“…Hybrid-TransCD [62] proposed a hybrid multi-scale transformer structure, which can model the mixed-scale attention representation of each image. STransUNet [66] combined transformer and UNet architectures, which can capture shallow detail features and model global context in high-level features. In order to capture the spatial and channel information of feature maps, MTCNet [67] divides CBAM into a SAM and a CAM, which are applied to the front-end and back-end of the multi-scale transformer, respectively.…”
Section: B Transformer-based Methodsmentioning
confidence: 99%
“…Hybrid-TransCD [62] proposed a hybrid multi-scale transformer structure, which can model the mixed-scale attention representation of each image. STransUNet [66] combined transformer and UNet architectures, which can capture shallow detail features and model global context in high-level features. In order to capture the spatial and channel information of feature maps, MTCNet [67] divides CBAM into a SAM and a CAM, which are applied to the front-end and back-end of the multi-scale transformer, respectively.…”
Section: B Transformer-based Methodsmentioning
confidence: 99%
“…To simplify the transformer block, Lin et al [44] presented the swin transformer framework by incorporating hierarchical, locality, and shift invariance priors, which achieves better performance in various tasks. Inspired by transformer, some CD methods based on transformer and CNN are proposed [45][46][47][48][49][50]. Shi et al [51] designed the bi-temporal image transformer (BIT), a method based on transformer to address CD challenges.…”
Section: Introductionmentioning
confidence: 99%
“…[44] presented the swin transformer framework by incorporating hierarchical, locality, and shift invariance priors, which achieves better performance in various tasks. Inspired by transformer, some CD methods based on transformer and CNN are proposed [45–50]. Shi et al.…”
Section: Introductionmentioning
confidence: 99%
“…In contrast, Transformer possesses prominent global modeling capability, and it is independent of the constraints of a traditional convolutional receptive field, thus making it more versatile and flexible [31]. However, a drawback of Transformer lies in its dependence on a greater number of deep feature sequences to indicate its performance, such that local fine granularity and the large computational parameter become difficult to perceive [32,33]. Thus, combining the strengths of CNN and Transformer in local and global modeling has effectively become an advanced design approach.…”
Section: Introductionmentioning
confidence: 99%