2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2021
DOI: 10.1109/cvprw53098.2021.00301
|View full text |Cite
|
Sign up to set email alerts
|

Distill on the Go: Online knowledge distillation in self-supervised learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
19
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 30 publications
(20 citation statements)
references
References 4 publications
1
19
0
Order By: Relevance
“…We train teacher and student networks with a batch size of 32 and 100 epochs. Our teacher model achieves a similar accuracy with literature [5] (52.02% vs. 48.26%). We train CVAE in the same way as in CIFAR-100.…”
Section: Results On Tiny-imagenetsupporting
confidence: 81%
“…We train teacher and student networks with a batch size of 32 and 100 epochs. Our teacher model achieves a similar accuracy with literature [5] (52.02% vs. 48.26%). We train CVAE in the same way as in CIFAR-100.…”
Section: Results On Tiny-imagenetsupporting
confidence: 81%
“…For example, [19] first uses self-distillation, which transfers the knowledge in the deeper portion of the networks to the shallow sections. [20] improves representation quality of the smaller models by single-stage online knowledge distillation. [21] uses self-distillation between different epochs with soft targets.…”
Section: Self-distillationmentioning
confidence: 99%
“…To ensure fairness of our comparisons, we apply the new training protocol to both the previous baseline AL models, and to our proposed DAES framework. The proposed training protocol could help to eventually unify this line of work with some form of knowledge distillation [46,47].…”
Section: Daes With Self-trained Knowledge Distillationmentioning
confidence: 99%