2023
DOI: 10.3390/s23073420
|View full text |Cite
|
Sign up to set email alerts
|

High-Resolution Swin Transformer for Automatic Medical Image Segmentation

Abstract: The resolution of feature maps is a critical factor for accurate medical image segmentation. Most of the existing Transformer-based networks for medical image segmentation adopt a U-Net-like architecture, which contains an encoder that converts the high-resolution input image into low-resolution feature maps using a sequence of Transformer blocks and a decoder that gradually generates high-resolution representations from low-resolution feature maps. However, the procedure of recovering high-resolution represen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 26 publications
(8 citation statements)
references
References 38 publications
0
8
0
Order By: Relevance
“…Refs. 14,16,27,30 integrated Transformer with other existing efficient frameworks. Specifically, Wei et al 14 proposed a HRSTNet which combined the HRNet and Swin Transformer.…”
Section: Other Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Refs. 14,16,27,30 integrated Transformer with other existing efficient frameworks. Specifically, Wei et al 14 proposed a HRSTNet which combined the HRNet and Swin Transformer.…”
Section: Other Methodsmentioning
confidence: 99%
“…14,16,27,30 integrated Transformer with other existing efficient frameworks. Specifically, Wei et al 14 proposed a HRSTNet which combined the HRNet and Swin Transformer. Li et al 16 adopted the structure of FC-DenseNet joined with ResLinear-Transformer (RL-Transformer) and convolutional linear attention block (CLAB) and proposed the TFCNs.…”
Section: Other Methodsmentioning
confidence: 99%
“…In recent years, due to the great achievements of deep learning in the field of medical image segmentation [ 21 , 22 , 23 ], researchers began to explore semi-automatic annotation based on deep learning [ 24 , 25 , 26 , 27 , 28 , 29 ]. Pair [ 30 ] is a semi-automatic annotation tool based on deep learning developed by Shenzhen University for medical image annotation tasks.…”
Section: Introductionmentioning
confidence: 99%
“…Notably, Swin Transformer (SwinT) offers scalability, allowing it to process images of different sizes and adapt to diverse datasets and resolutions while maintaining competitive performance, reducing computational demands [ 21 ]. Impressively, SwinT approaches exhibit robust generalization capabilities across a spectrum of computer vision tasks with minimal task-specific adjustments, and their interpretability and mitigation of overfitting contribute to their appeal [ 22 , 23 ].…”
Section: Introductionmentioning
confidence: 99%