2018
DOI: 10.48550/arxiv.1807.02802
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Revisiting Distillation and Incremental Classifier Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 0 publications
0
4
0
Order By: Relevance
“…We compare with the well-known rehearsal-based method iCaRL which proposes to select representative samples for old class incremental tasks and use a distillation loss to train. In our reimplementation, we set the capacity of representative set K = 2000 and use random exemplar selection rather than exemplar selection by herding since there is no substantial difference as claimed in [43]. And we train our ICLNet with the same selected representative sample set for fair comparison.…”
Section: F Compare With Rehearsal-based Methodsmentioning
confidence: 99%
“…We compare with the well-known rehearsal-based method iCaRL which proposes to select representative samples for old class incremental tasks and use a distillation loss to train. In our reimplementation, we set the capacity of representative set K = 2000 and use random exemplar selection rather than exemplar selection by herding since there is no substantial difference as claimed in [43]. And we train our ICLNet with the same selected representative sample set for fair comparison.…”
Section: F Compare With Rehearsal-based Methodsmentioning
confidence: 99%
“…We argue that pruning secondary parameters is sub-optimal in the case of single-head protocol Memory rehearsal and pseudo memory rehearsal. To mitigate knowledge bias toward new tasks, some methods store previous data and retrain them [5,7,8,21,22], or train generative adversarial networks (GANs) to generate and discriminate images and then learn the data distribution [28][29][30][31]. Memory rehearsal methods require additional storage to store previous data or extra model parameters to generate and discriminate data.…”
Section: Related Workmentioning
confidence: 99%
“…In the family of the single-network methods, previous works have explored the regularization methods [2,4,[16][17][18], the parameter isolation methods [19,20] and the memory rehearsal methods [5,7,8,21,22]. The regularization methods leverage a penalty term in the loss function to regularize the parameters when updating for new tasks.…”
Section: Introductionmentioning
confidence: 99%
“…In the family of the single-network methods, previous works have explored the regularization methods [1,4,14,18,33], the parameter isolation methods [21,22] and the memory rehearsal methods [3,5,12,20,27]. The regularization methods leverage a penalty term in the loss function to regularize the parameters when updating for new tasks.…”
Section: Introductionmentioning
confidence: 99%