2021
DOI: 10.1007/s00521-021-05981-0
|View full text |Cite
|
Sign up to set email alerts
|

Effective training of convolutional neural networks for age estimation based on knowledge distillation

Abstract: Age estimation from face images can be profitably employed in several applications, ranging from digital signage to social robotics, from business intelligence to access control. Only in recent years, the advent of deep learning allowed for the design of extremely accurate methods based on convolutional neural networks (CNNs) that achieve a remarkable performance in various face analysis tasks. However, these networks are not always applicable in real scenarios, due to both time and resource constraints that t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 24 publications
(9 citation statements)
references
References 51 publications
0
9
0
Order By: Relevance
“…Knowledge distillation has been shown to improve robustness to image corruptions [38,40,42,18]. Here, we demonstrate, for the first time, that a V1-inspired CNN can lead to higher robustness gains through distillation.…”
Section: Discussionmentioning
confidence: 56%
See 3 more Smart Citations
“…Knowledge distillation has been shown to improve robustness to image corruptions [38,40,42,18]. Here, we demonstrate, for the first time, that a V1-inspired CNN can lead to higher robustness gains through distillation.…”
Section: Discussionmentioning
confidence: 56%
“…Ensembling is a well-known machine learning technique to combine smaller individual models into a larger model leading to superior performance compared to the individual models in a diverse range of supervised learning problems [24,25,26,27,28,29], including generalization to out-of-domain (OOD) datasets [30,31,32,33,34]. Knowledge Distillation is a popular technique [35,36,37,38,39,40,41] to transfer the superior performance of the ensemble (teacher) into a single smaller model (student). This technique has been shown to improve the generalization ability of the student models [38,40,42,18].…”
Section: Ensemble and Distillationmentioning
confidence: 99%
See 2 more Smart Citations
“…The proposed model is relatively compact and suitable for use on mobile and embedded devices. Greco et al [35] propose an effective training method for age estimation CNNs based on knowledge distillation. The goal of the method is to first learn age estimation using a more complex neural network and then distill the knowledge to a smaller, more compact model.…”
Section: Related Workmentioning
confidence: 99%