2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2022
DOI: 10.1109/cvprw56347.2022.00309
|View full text |Cite
|
Sign up to set email alerts
|

ResNeSt: Split-Attention Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
303
0
2

Year Published

2022
2022
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 750 publications
(308 citation statements)
references
References 38 publications
3
303
0
2
Order By: Relevance
“…e residual block is very efficient to build a deeper network. More recently, ResNeXt [22] and ResNest [23] have been designed to improve feature extraction process. Although these above deep CNN architectures extract high-quality object features, they require intensive computations due to large numbers of parameters.…”
Section: Lightweight Network For Extracting High-level Semanticmentioning
confidence: 99%
“…e residual block is very efficient to build a deeper network. More recently, ResNeXt [22] and ResNest [23] have been designed to improve feature extraction process. Although these above deep CNN architectures extract high-quality object features, they require intensive computations due to large numbers of parameters.…”
Section: Lightweight Network For Extracting High-level Semanticmentioning
confidence: 99%
“…The approach can be applied in principle to any CNN. In the following experiments we have tested VGG16, as it was commonly used in many competing methods and comparisons, ResNet, as it has been used in the most recent state-of-the-art methods, and the novel ResNeSt architecture [58] that has recently improved results over previous ResNet in image classification, object detection, instance segmentation and semantic segmentation.…”
Section: Pooling Of Local Cnn Features and Descriptor Splittingmentioning
confidence: 99%
“…On these premises, and after selecting our more prominent setup (4-Split + ReLU + Max-Poolings), we proceed for a more time-expensive experiment training the networks at resolution 512 × 512. Finally, we test a very recent ResNet variant, the ResNeSt101 and ResNeSt50 [58], that have shown improved results in different tasks, from segmentation to classification; our experiments show that this architecture heavily outperforms its more traditional version also for the image retrieval task. In our experimental setup, the ResNeSt101 and ResNeSt50, respectively, completed the features extraction from the test datasets in about 120% and 85% of the total time employed by the original ResNet101.…”
Section: Trainings At Higher Resolution With Modern Architecturesmentioning
confidence: 99%
See 1 more Smart Citation
“…The backbone network of the semantic feature enhancement module is the ResNest [27] model. The Split‐Attention module in the network groups the channel dimensions of the feature map and calculates the attention weights corresponding to each group of features, which effectively improves the quality of features extracted from the model without significantly increasing the model computation and parameters.…”
Section: Approachmentioning
confidence: 99%