2022
DOI: 10.1007/978-3-031-16437-8_11
|View full text |Cite
|
Sign up to set email alerts
|

Stepwise Feature Fusion: Local Guides Global

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
48
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 128 publications
(48 citation statements)
references
References 18 publications
0
48
0
Order By: Relevance
“…SSFormer [3], which uses a pyramid transformer encoder and a progressive locality decoder for stepwise feature fusion with local emphasis was trained as a reference method. This state-of-the-art model efficiently segments complex morphological features of varying sizes.…”
Section: Methodsmentioning
confidence: 99%
“…SSFormer [3], which uses a pyramid transformer encoder and a progressive locality decoder for stepwise feature fusion with local emphasis was trained as a reference method. This state-of-the-art model efficiently segments complex morphological features of varying sizes.…”
Section: Methodsmentioning
confidence: 99%
“…For example, MIA‐Net [8] uses both to capture global dependences and low‐level spatial details. Some studies [4, 6, 7, 9, 54] adopt pure Transformers for feature abstraction. For example, ColonFormer [6] and Polyp2Seg [7] rely on Transformer encoder for feature encoding and atrous convolution to further explore multi‐scale representations.…”
Section: Related Workmentioning
confidence: 99%
“…This characteristic fits well with the advantages of Transformers which do well in discovering the long‐range relations. Actually, the encoders built by Transformer [5], or Transformer encoders, have already been adopted in recent studies [4, 69] for robust feature extraction when doing polyp segmentation, which is also interesting to us for better performance.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Zhou et al 23 also achieved improved accuracy on the BCTV dataset by replacing most of the convolutional blocks in Vnet with a 3D sliding window ViT model that implemented a 3D extension of efficient Swin-like transformer layers. 24 Not limited to the networks mentioned above, more transformerbased segmentation networks [25][26][27][28][29][30] were proposed and demonstrated state-of -the-art performance in different medical segmentation tasks.…”
Section: Introductionmentioning
confidence: 99%