2018 25th IEEE International Conference on Image Processing (ICIP) 2018
DOI: 10.1109/icip.2018.8451808
|View full text |Cite
|
Sign up to set email alerts
|

Discover the Effective Strategy for Face Recognition Model Compression by Improved Knowledge Distillation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(9 citation statements)
references
References 5 publications
0
9
0
Order By: Relevance
“…In last few years, a variety of knowledge distillation methods have been widely used for model compression in different visual recognition applications. Specifically, most of the knowledge distillation methods were previously developed for image classification (Li and Hoiem, 2017;Peng et al, 2019b;Bagherinezhad et al, 2018;Chen et al, 2018a;Wang et al, 2019b;Mukherjee et al, 2019;Zhu et al, 2019) and then extended to other visual recognition applications, including face recognition (Luo et al, 2016;Kong et al, 2019;Yan et al, 2019;Ge et al, 2018;Wang et al, 2018bWang et al, , 2019cDuong et al, 2019;Wu et al, 2020;Wang et al, 2017), action recognition (Hao and Zhang, 2019;Thoker and Gall, 2019;Luo et al, 2018;Garcia et al, 2018;Wu et al, 2019b;Zhang et al, 2020), object detection Hong and Yu, 2019;Shmelkov et al, 2017;Wei et al, 2018;Wang et al, 2019d), lane detection (Hou et al, 2019), image or video segmentation (He et al, 2019;Liu et al, 2019g;Mullapudi et al, 2019;Siam et al, 2019;Dou et al, 2020), video classification (Bhardwaj et al, 2019;Zhang and Peng, 2018), pedestrian detection (Shen et al, 2016), facial landmark detection (Dong and Yang, 2019), person re-identification (Wu et al, 2019a)…”
Section: Kd In Visual Recognitionmentioning
confidence: 99%
See 1 more Smart Citation
“…In last few years, a variety of knowledge distillation methods have been widely used for model compression in different visual recognition applications. Specifically, most of the knowledge distillation methods were previously developed for image classification (Li and Hoiem, 2017;Peng et al, 2019b;Bagherinezhad et al, 2018;Chen et al, 2018a;Wang et al, 2019b;Mukherjee et al, 2019;Zhu et al, 2019) and then extended to other visual recognition applications, including face recognition (Luo et al, 2016;Kong et al, 2019;Yan et al, 2019;Ge et al, 2018;Wang et al, 2018bWang et al, , 2019cDuong et al, 2019;Wu et al, 2020;Wang et al, 2017), action recognition (Hao and Zhang, 2019;Thoker and Gall, 2019;Luo et al, 2018;Garcia et al, 2018;Wu et al, 2019b;Zhang et al, 2020), object detection Hong and Yu, 2019;Shmelkov et al, 2017;Wei et al, 2018;Wang et al, 2019d), lane detection (Hou et al, 2019), image or video segmentation (He et al, 2019;Liu et al, 2019g;Mullapudi et al, 2019;Siam et al, 2019;Dou et al, 2020), video classification (Bhardwaj et al, 2019;Zhang and Peng, 2018), pedestrian detection (Shen et al, 2016), facial landmark detection (Dong and Yang, 2019), person re-identification (Wu et al, 2019a)…”
Section: Kd In Visual Recognitionmentioning
confidence: 99%
“…The existing KD-based face recognition methods can not only be easily deployed, but also improve the classification accuracy (Luo et al, 2016;Kong et al, 2019;Yan et al, 2019;Ge et al, 2018;Wang et al, 2018bWang et al, , 2019cDuong et al, 2019;Wang et al, 2017). First of all, these methods focus on the lightweight face recognition with very satisfactory accuracy (Luo et al, 2016;Wang et al, 2018bWang et al, , 2019cDuong et al, 2019). In (Luo et al, 2016), the knowledge from the chosen informative neurons of top hint layer of the teacher network is transferred into the student network.…”
Section: Kd In Visual Recognitionmentioning
confidence: 99%
“…For the later, we use L2 loss for features or normalized features to make a constraint. For feature normalizing, the features are normalized as x/x to make them have the same scale, and with [47], the optimization of normalized feature L2 loss LfeatNorm becomes consistent with cosine similarity compared with the feature L2 loss Lfeat, which can make an performance improvement. In all evaluations, only the identity features extracted from low-quality images are used to match.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…In this context, distillation and pruning techniques, exploring the deep model parameters redundancy to remove uncritical information, can be adopted to perform model compression and acceleration, without significantly reducing the recognition performance [103, 104]. This may facilitate the deployment of deep face models on resource‐constrained embedded devices such as mobile devices [105]. Face presentation attack detection for face recognition security – Despite the significant progress in face recognition performance, the widespread adoption of face recognition solutions raises new security concerns.…”
Section: Challenges and Future Research Directionsmentioning
confidence: 99%
“…In this context, distillation and pruning techniques, exploring the deep model parameters redundancy to remove uncritical information, can be adopted to perform model compression and acceleration, without significantly reducing the recognition performance [103, 104]. This may facilitate the deployment of deep face models on resource‐constrained embedded devices such as mobile devices [105].…”
Section: Challenges and Future Research Directionsmentioning
confidence: 99%