2022
DOI: 10.1155/2022/6174255
|View full text |Cite
|
Sign up to set email alerts
|

A Defect Detection Model for Industrial Products Based on Attention and Knowledge Distillation

Abstract: Industrial quality detection is one of the important fields in machine vision. Big data analysis, the Internet of Things, edge computing, and other technologies are widely used in industrial quality detection. Studying an industrial detection algorithm that can be organically combined with the Internet of Things and edge computing is imminent. Deep learning methods in industrial quality detection have been widely proposed recently. However, due to the particularity of industrial scenarios, the existing deep le… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 33 publications
0
3
0
Order By: Relevance
“…Considering the shortcomings related to the use of transformers in the industrial defect recognition field, the authors can find different transformer-based approaches applied to the defect detection task. Both Gao et al [283] and Zhang et al [269] proposed a swin-transformer model. The former designed a new window-shift scheme that further strengthened the feature transfer between the windows.…”
Section: E: Exploiting Transformersmentioning
confidence: 99%
See 1 more Smart Citation
“…Considering the shortcomings related to the use of transformers in the industrial defect recognition field, the authors can find different transformer-based approaches applied to the defect detection task. Both Gao et al [283] and Zhang et al [269] proposed a swin-transformer model. The former designed a new window-shift scheme that further strengthened the feature transfer between the windows.…”
Section: E: Exploiting Transformersmentioning
confidence: 99%
“…The pseudo-labels are the output probabilities from the teacher and these concur to guide the student learning process along with the hard (i.e., real) labels. During training, the student model endeavours to match the teacher output probabilities regulated by the shared weights; thus, a ''distillation loss'', sometimes assisted by an attention module [269], steers the probability distribution generated by the student model towards that provided by the teacher model. The loss minimization process can be related to a network-level, whenever the knowledge is optimized only by the last layers or to a channel-level if knowledge is optimized at different levels of feature maps, de facto computing multiple losses at time.…”
Section: ) Knowledge Distillationmentioning
confidence: 99%
“…In the field of industrial surface defect detection, Li et al [39] proposed an attention enhancement-based PANet feature fusion method. Zhang et al [40] proposed a lightweight industrial training model based on knowledge distillation while also incorporating the transformer self-attention used to extract global features. Wang et al [41] explored different types of shallow and lightweight neural networks, including supervised and unsupervised models, to improve the fall detection results.…”
Section: Surface Defect Detectionmentioning
confidence: 99%