2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412615
|View full text |Cite
|
Sign up to set email alerts
|

Feature Fusion for Online Mutual Knowledge Distillation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
34
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 56 publications
(34 citation statements)
references
References 11 publications
0
34
0
Order By: Relevance
“…Recently, there are many new online distillation methods that train a teacher and a student simultaneously during knowledge transfer. Collaborative learning is the one used most often [25,29,32,8,9,7], where the teacher and the student as peer networks collaboratively teach and learn from each other, and the peer network architectures can be different. In particular, Zhang et al [25] proposed a deep mutual learning method (DML) for online distillation using the response-based knowledge.…”
Section: Collaborative Learningmentioning
confidence: 99%
See 3 more Smart Citations
“…Recently, there are many new online distillation methods that train a teacher and a student simultaneously during knowledge transfer. Collaborative learning is the one used most often [25,29,32,8,9,7], where the teacher and the student as peer networks collaboratively teach and learn from each other, and the peer network architectures can be different. In particular, Zhang et al [25] proposed a deep mutual learning method (DML) for online distillation using the response-based knowledge.…”
Section: Collaborative Learningmentioning
confidence: 99%
“…Unlike the ensemble of peer networks, the advantage of a mutual knowledge distillation method is that it can fuse features of peer networks to collaboratively learn a powerful classifier [32]. However, the knowledge distilled by those online mutual distillation methods is limited to the response-based knowledge from individual instance features.…”
Section: Collaborative Learningmentioning
confidence: 99%
See 2 more Smart Citations
“…Teacher network transfers its knowledge to student network to enhance the performance of student network. Feature maps [8,9,10] and logits of a network [11,12] are widely used as knowledge. Model compression has actively been studied mainly on computer vision tasks.…”
Section: Introductionmentioning
confidence: 99%