2021 43rd Annual International Conference of the IEEE Engineering in Medicine &Amp; Biology Society (EMBC) 2021
DOI: 10.1109/embc46164.2021.9630336
|View full text |Cite
|
Sign up to set email alerts
|

UCATR: Based on CNN and Transformer Encoding and Cross-Attention Decoding for Lesion Segmentation of Acute Ischemic Stroke in Non-contrast Computed Tomography Images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(9 citation statements)
references
References 11 publications
0
9
0
Order By: Relevance
“…The images are learnt for a global representation. The hierarchical Swin Transformer with an offset window as an encoder to represent the image spatial context features and design a decoder with a patch extended layer for upsampling operations to recover the spatial resolution of the feature graph [21] . SWin-Unet's architecture is shown in Figure 2.…”
Section: Swin-unet: Unet-like Pure Transformer For Tooth Segmentationmentioning
confidence: 99%
See 1 more Smart Citation
“…The images are learnt for a global representation. The hierarchical Swin Transformer with an offset window as an encoder to represent the image spatial context features and design a decoder with a patch extended layer for upsampling operations to recover the spatial resolution of the feature graph [21] . SWin-Unet's architecture is shown in Figure 2.…”
Section: Swin-unet: Unet-like Pure Transformer For Tooth Segmentationmentioning
confidence: 99%
“…Traditional CNN shows good performance in image segmentation, but due to the limitation of convolutional operations, the global and remote semantic information interaction cannot be learned well. SWin-Unet builds a symmetric encoder-decoder structure with jump connections, implements a local to global self-attention mechanism, and develops a patch expanding layer without convolution and interpolation operations (patch expanding layer) to increases upsampling and feature dimension, which outperforms convolution based methods [20,21] .…”
Section: Introductionmentioning
confidence: 99%
“…Based on the U-Net structure, Cao et al progressed the SWIN-Unit transformer algorithm, superior to U-Net and other models in multiorgan and heart segmentation tasks, with a DSC value of 0.9 [ 52 ]. In ischemic stroke lesion analysis, the model, including CNN and transformer for encoding and the multihead cross-attention (MHCA) module for decoding, leads to stroke lesion morphology and edges with a Dice of 73.58% [ 53 ]. Transformers can be introduced to process data with different scanner models or multimodal data.…”
Section: Deep Learning Modelsmentioning
confidence: 99%
“…The skip connection in the developed U-shape model is composed of a residual atrous pyramid pooling module, a position-sensitive axial attention module, and a hierarchical context-attention module. Luo and colleagues [148] designed a model named UCATR to segment acute ischemic stroke lesions using the NCCT CT dataset. The proposed model is composed of a CNN-transformer encoder and a cross-attention decoder.…”
Section: Segmentationmentioning
confidence: 99%