2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00297
|View full text |Cite
|
Sign up to set email alerts
|

Snapshot Distillation: Teacher-Student Optimization in One Generation

Abstract: Optimizing a deep neural network is a fundamental task in computer vision, yet direct training methods often suffer from over-fitting. Teacher-student optimization aims at providing complementary cues from a model trained previously, but these approaches are often considerably slow due to the pipeline of training a few generations in sequence, i.e., time complexity is increased by several times. This paper presents snapshot distillation (SD), the first framework which enables teacher-student optimization in on… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
115
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 177 publications
(115 citation statements)
references
References 45 publications
0
115
0
Order By: Relevance
“…1, compared with iTracker [9], our result outperforms in both iphone dataset and ipad dataset, achieving 4.8% and 5.3% error reduction respectively. Compared with SD [26] which is reproduced by us for gaze estimation, our error is a little higher in ipad set, but lower in iphone set and the total set. iphone ipad total iTracker [9] 1.86 2.81 2.05 SD [26] 1.81 2.61 1.97 TAT 1.77 2.66 1.95 Table 1.…”
Section: Comparison With State Of the Artmentioning
confidence: 55%
See 4 more Smart Citations
“…1, compared with iTracker [9], our result outperforms in both iphone dataset and ipad dataset, achieving 4.8% and 5.3% error reduction respectively. Compared with SD [26] which is reproduced by us for gaze estimation, our error is a little higher in ipad set, but lower in iphone set and the total set. iphone ipad total iTracker [9] 1.86 2.81 2.05 SD [26] 1.81 2.61 1.97 TAT 1.77 2.66 1.95 Table 1.…”
Section: Comparison With State Of the Artmentioning
confidence: 55%
“…After that, we compare with two methods. The first one is iTracker [9], which is the state-of-the-art in 2-D gaze estimation, and the second one is SD [26], which is the state-of-the-art general knowledge distillation method. We reproduce their method for gaze estimation task.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations