2021
DOI: 10.1007/978-3-030-87589-3_40
|View full text |Cite
|
Sign up to set email alerts
|

GT U-Net: A U-Net Like Group Transformer Network for Tooth Root Segmentation

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4
1

Relationship

2
8

Authors

Journals

citations
Cited by 40 publications
(22 citation statements)
references
References 21 publications
1
21
0
Order By: Relevance
“…Inspired by this, Radford et al proposed CLIP [17], which uses contrast learning to learn image representations on a dataset of 400 million pairs (image, text) from scratch. And in the natural image semantic segmentation, there are studies [18][19][20] that have begun to use text information to improve the segmentation capabilities of models. In the field of medical image analysis, there are some related works too [21,22].…”
Section: Related Workmentioning
confidence: 99%
“…Inspired by this, Radford et al proposed CLIP [17], which uses contrast learning to learn image representations on a dataset of 400 million pairs (image, text) from scratch. And in the natural image semantic segmentation, there are studies [18][19][20] that have begun to use text information to improve the segmentation capabilities of models. In the field of medical image analysis, there are some related works too [21,22].…”
Section: Related Workmentioning
confidence: 99%
“…They used an improved region growing algorithm and parameter adaptive method to expand the resemble regions and remove unnecessary parameters to enhance their segmentation performance. To cost-effectively improve the results of tooth segmentation, Li et al (2021) proposed a group transformer to achieve advanced performance on tooth root segmentation. Koch et al (2019) input the original image into the network in blocks and achieved pretty performance through U-Net.…”
Section: Related Workmentioning
confidence: 99%
“…The first one, which is represented by ViT [11], is to take a patch with a certain size as an element, but it is in the image level and more suitable for classification tasks. The second one is Swin Transformer [12,13,14], which performs pixel-level operations in a small local window, and gradually obtains global information via deepening the network, but it does not fully utilize Transformer's global modeling capability. Therefore, we design a novel model so-called Shuffle Transformer to capture global capability with low computational complexity.…”
Section: Shape Autoencoder (Sae)mentioning
confidence: 99%