2022
DOI: 10.1007/978-3-031-09002-8_21
|View full text |Cite
|
Sign up to set email alerts
|

Combining CNNs with Transformer for Multimodal 3D MRI Brain Tumor Segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 11 publications
0
2
0
Order By: Relevance
“…In addition, the spatial attention mechanism allowed the model to extract the important regions of the image adaptively. Different from the disadvantage of the poor data correlation caused by dividing the feature map, the feature map could be directly converted into tokens by spatial pooling without segmentation, which effectively ensured the consistency of feature information [ 80 , 81 , 82 ]. In this paper, six groups of spatial attention mechanisms were set for each Transformer module to extract the network, i.e., the number of tokens input to each Transformer module was four.…”
Section: Methodsmentioning
confidence: 99%
“…In addition, the spatial attention mechanism allowed the model to extract the important regions of the image adaptively. Different from the disadvantage of the poor data correlation caused by dividing the feature map, the feature map could be directly converted into tokens by spatial pooling without segmentation, which effectively ensured the consistency of feature information [ 80 , 81 , 82 ]. In this paper, six groups of spatial attention mechanisms were set for each Transformer module to extract the network, i.e., the number of tokens input to each Transformer module was four.…”
Section: Methodsmentioning
confidence: 99%
“…The recent success of transformer architecture in vision tasks ( 11 , 12 ) has shown benefits in learning global contextual information. New network designs with vision transformers have emerged for medical image segmentation ( 13 , 14 ) and achieved state-of-the-art (SOTA) performance in brain tumor segmentation ( 15 17 ). However, the supervised training of vision transformers typically requires a large amount of densely annotated images, otherwise there is a high risk of overfitting.…”
Section: Introductionmentioning
confidence: 99%