2023
DOI: 10.1016/j.sysarc.2023.102861
|View full text |Cite
|
Sign up to set email alerts
|

Anomaly detection based on multi-teacher knowledge distillation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(2 citation statements)
references
References 21 publications
0
2
0
Order By: Relevance
“…The intuition of KD is to minimize the divergence between logits outputs from teacher model and student model. Therefore, student model can learn the knowledge from teacher model and acquire a better performance because student model could approach output from teacher model 30 , 31 . The loss of knowledge distillation could be computed by Kullback-Leiler divergence: Based on the type of transferred knowledge, knowledge distillation can be divided into three categories: (1) output transfer; (2) feature transfer; (3) relation transfer.…”
Section: Background and Related Workmentioning
confidence: 99%
“…The intuition of KD is to minimize the divergence between logits outputs from teacher model and student model. Therefore, student model can learn the knowledge from teacher model and acquire a better performance because student model could approach output from teacher model 30 , 31 . The loss of knowledge distillation could be computed by Kullback-Leiler divergence: Based on the type of transferred knowledge, knowledge distillation can be divided into three categories: (1) output transfer; (2) feature transfer; (3) relation transfer.…”
Section: Background and Related Workmentioning
confidence: 99%
“…The intuition of KD is to minimize the divergence between logits outputs from teacher model and student model. Therefore, student model can learn the knowledge from teacher model and acquire a better performance because student model could approach output from teacher model 27,28 . The loss of knowledge distillation could be computed by Kullback-Leiler divergence:…”
Section: Knowledge Distillationmentioning
confidence: 99%