2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01031
|View full text |Cite
|
Sign up to set email alerts
|

Neural Architecture Search for Lightweight Non-Local Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
34
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 54 publications
(34 citation statements)
references
References 21 publications
0
34
0
Order By: Relevance
“…As for our future work, we will strive for the early application of the proposed method to clinical practice. First, we will further explore the potential of the proposed method with other efficient architectures, for instance, EfficientNet [38], Non-local-Block [39]; optimize the hyper-parameters of the objective and teste the applicability of our method in variant clinical datasets. Second, we will extend our method into multidomain transfer, which is still a challenge and very important in practice.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…As for our future work, we will strive for the early application of the proposed method to clinical practice. First, we will further explore the potential of the proposed method with other efficient architectures, for instance, EfficientNet [38], Non-local-Block [39]; optimize the hyper-parameters of the objective and teste the applicability of our method in variant clinical datasets. Second, we will extend our method into multidomain transfer, which is still a challenge and very important in practice.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…[27,34,48] adapts the idea of BERT to vision and language tasks and showed improved performance on multiple sub-tasks. [55] bridges attention and non-local operator to capture longrange dependency, which has been used in many computer vision tasks [67,28,5,64]. In our work, we apply attention over a group of images and show its effectiveness for summarizing information in an image group.…”
Section: Related Workmentioning
confidence: 99%
“…These networks aim to capture global dependencies in images and videos. Note that non-local neural networks are inspired by the classical non-local method in vision tasks [3] and unlike those in Transformers, the selfattentions in non-local networks are usually not equipped with multi-heads and position embedding [37,4,23]. Afterwards, Transformers achieve remarkable success in NLP tasks [12,25] and, therefore, self-attentions that inherits NLP settings (e.g., multi-heads, position encodings, classification token, etc.)…”
Section: Related Workmentioning
confidence: 99%