2022
DOI: 10.48550/arxiv.2201.12785
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

TransBTSV2: Towards Better and More Efficient Volumetric Segmentation of Medical Images

Abstract: Transformer, benefiting from global (longrange) information modeling using self-attention mechanism, has been successful in natural language processing and computer vision recently. Convolutional Neural Networks, capable of capturing local features, are unable to model explicit long-distance dependencies from global feature space. However, both local and global features are crucial for dense prediction tasks, especially for 3D medical image segmentation. In this paper, we exploit Transformer in 3D CNN for 3D m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 31 publications
0
4
0
Order By: Relevance
“…We compare our model with the state-of-the-art models on the BraTS 2020 validation set, with experimental data from Li et al (32). The experimental data shows that our model obtains the best results on ET and WT (Table 2).…”
Section: Experiments Resultsmentioning
confidence: 99%
“…We compare our model with the state-of-the-art models on the BraTS 2020 validation set, with experimental data from Li et al (32). The experimental data shows that our model obtains the best results on ET and WT (Table 2).…”
Section: Experiments Resultsmentioning
confidence: 99%
“…Through the feature fusion and analysis from the Transformer and CNN, TransBTS produced high-resolution segmentation. Subsequently, an improved version of TransBTS called TransBTSV2 was developed, [129] which contained a variable bottleneck module with the Transformer for shape and detail awareness, thereby achieving superior performance. Lyu et al [130] also replaced the bottleneck layer with 12 consecutive Transformers to obtain attention enhancement.…”
Section: Brain Tumor Segmentationmentioning
confidence: 99%
“…In this regard, TransBTS [10] network is proposed, which incorporates Transformer into the 3D CNN encoder-decoder architecture for the first time, enhancing global feature extraction. TransBTSV2 [11] further improves TransBTS by redesigning the Transformer module www.ijacsa.thesai.org and introducing deformable bottleneck module to capture shape-sensitive local features. SwinBTS [12] structure employs 3D SwinTransformer as both encoder and decoder of the network to extract global information from feature maps efficiently, using convolution operation for upsampling and downsampling.…”
Section: Related Workmentioning
confidence: 99%