2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.01621
|View full text |Cite
|
Sign up to set email alerts
|

General Multi-label Image Classification with Transformers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
94
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 239 publications
(94 citation statements)
references
References 28 publications
0
94
0
Order By: Relevance
“…As it develops, it has almost replaced RNN with its excellent effects, such as exploring long-range relationship, better scalability to highcomplexity models and so on. We can observe that transformer has been applied to image classification (Lanchantin et al 2021;Dosovitskiy et al 2020), object detection (Yang et al 2020;Zhu et al 2021;Carion et al 2020) and even natural image restoration (Wang et al 2021;Liang et al 2021).…”
Section: Transformermentioning
confidence: 99%
“…As it develops, it has almost replaced RNN with its excellent effects, such as exploring long-range relationship, better scalability to highcomplexity models and so on. We can observe that transformer has been applied to image classification (Lanchantin et al 2021;Dosovitskiy et al 2020), object detection (Yang et al 2020;Zhu et al 2021;Carion et al 2020) and even natural image restoration (Wang et al 2021;Liang et al 2021).…”
Section: Transformermentioning
confidence: 99%
“…We inject symmetric label noise (details in Section 4.1) at various corruption levels, from 0% to 60%, and report the mean average precision to asses the impact of wrong labels. mAP is considered by many recent works [Lanchantin et al, 2020, Chen et al, 2019b as important metric for performance evaluation in multi-label classification, since it takes into consideration both false-negative and false-positive rates [Baruch et al, 2020]. Fig.…”
Section: Motivation Examplementioning
confidence: 99%
“…In addition, we also exploit mask token strategy [7,25] to further strengthen the learned representation in the auxiliary task. During the training phase, we randomly set the state of domain information as either positive or negative instead of masking them out as unknown.…”
Section: Introductionmentioning
confidence: 99%