2022
DOI: 10.1016/j.compbiomed.2022.105954
|View full text |Cite
|
Sign up to set email alerts
|

SWTRU: Star-shaped Window Transformer Reinforced U-Net for medical image segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
13
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 27 publications
(13 citation statements)
references
References 9 publications
0
13
0
Order By: Relevance
“…As an alternative form of deep learning, transformers have already performed well in natural language processing 44 . In computer vision, transformer‐based architectures have also outperformed CNN‐based architectures in many classification tasks 45–47 . CNN is inherently designed to exploit the spatial locality in images due to the local‐receptive fields of convolutional filters.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…As an alternative form of deep learning, transformers have already performed well in natural language processing 44 . In computer vision, transformer‐based architectures have also outperformed CNN‐based architectures in many classification tasks 45–47 . CNN is inherently designed to exploit the spatial locality in images due to the local‐receptive fields of convolutional filters.…”
Section: Methodsmentioning
confidence: 99%
“…44 In computer vision, transformer-based architectures have also outperformed CNN-based architectures in many classification tasks. [45][46][47] CNN is inherently designed to exploit the spatial locality in images due to the local-receptive fields of convolutional filters. Convolutional layers scan the input image with window filters, enabling the model to learn hierarchical representations of local patterns.…”
Section: Vision Transformermentioning
confidence: 99%
“…This outcome pushed the number of layers in the neural network to 201, as shown in figure 2(a).The transformer(Vaswani et al 2017), another deep learning architecture originally employed for natural language processing, was recently developed into the vision transformer (ViT). In experiments, ViT indeed achieved higher accuracy than ResNet(Sengar et al 2022, Xin et al 2022, Zhang et al 2022b. ViT divides images into a series of patches to consider global correlations in feature extraction(Dosovitskiy et al 2020).…”
mentioning
confidence: 92%
“…Affected by the great success of U‐Net in medical image segmentation, U‐Net++ 17 and ResUNet++ 18 are also suitable for medical image segmentation and have achieved good results. Nowadays, most researchers achieve the purpose of improving network performance by adding or modifying various modules 19–25 . PraNet 19 proposed a parallel reverse attention network structure for segmentation using the polyp area and boundary information as cues.…”
Section: Introductionmentioning
confidence: 99%
“…Nowadays, most researchers achieve the purpose of improving network performance by adding or modifying various modules. [19][20][21][22][23][24][25] PraNet 19 proposed a parallel reverse attention network structure for segmentation using the polyp area and boundary information as cues. And aggregates the features with high-level semantics using a parallel partial decoder (PPD).…”
mentioning
confidence: 99%