2022
DOI: 10.48550/arxiv.2201.06945
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

It's All in the Head: Representation Knowledge Distillation through Classifier Sharing

Abstract: Representation knowledge distillation aims at transferring rich information from one model to another. Current approaches for representation distillation mainly focus on the direct minimization of distance metrics between the models' embedding vectors. Such direct methods may be limited in transferring high-order dependencies embedded in the representation vectors, or in handling the capacity gap between the teacher and student models. In this paper, we introduce two approaches for enhancing representation dis… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 1 publication
0
1
0
Order By: Relevance
“…Ben-Baruch et al [25] trained MobileNet using Knowledge Distillation to meet the demand for lightweight models. Alansari et al [26] presented a face feature extraction network based on GhostNet, which linearly replicates duplicated features, as another study on lightweight models.…”
Section: B Facial Recognition 1) General Facial Recognitionmentioning
confidence: 99%
“…Ben-Baruch et al [25] trained MobileNet using Knowledge Distillation to meet the demand for lightweight models. Alansari et al [26] presented a face feature extraction network based on GhostNet, which linearly replicates duplicated features, as another study on lightweight models.…”
Section: B Facial Recognition 1) General Facial Recognitionmentioning
confidence: 99%