2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00460
|View full text |Cite
|
Sign up to set email alerts
|

Focal and Global Knowledge Distillation for Detectors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
61
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 187 publications
(61 citation statements)
references
References 22 publications
0
61
0
Order By: Relevance
“…Due to the large-scale variation of objects in industrial detection datasets and the uneven distribution of positive and negative samples of objects, these will adversely affect the distillation effect. For this reason, this study uses the scaling mask to balance the object scale with reference to FGD [ 44 ] to solve the problem of an unbalanced object scale. The formula is as follows: where GT is the region of the ground truth.…”
Section: Methodsmentioning
confidence: 99%
“…Due to the large-scale variation of objects in industrial detection datasets and the uneven distribution of positive and negative samples of objects, these will adversely affect the distillation effect. For this reason, this study uses the scaling mask to balance the object scale with reference to FGD [ 44 ] to solve the problem of an unbalanced object scale. The formula is as follows: where GT is the region of the ground truth.…”
Section: Methodsmentioning
confidence: 99%
“…Dai et al [26] developed GID framework which selects distillation areas based on differences between the student and teacher networks. Yang et al proposed FGD [5] which separates the foreground and background, enabling the student model to learn from the teacher network areas of interest and global knowledge via local and global distillation respectively. Besides, MGD [6] imposes random masking on the feature map of the student model, and then generates the feature map reconstructing from the teacher network.…”
Section: B Knowledge Distillationmentioning
confidence: 99%
“…FGFI [4] operates by distilling the features of high IoU between ground truth and anchors. FGD [5] was developed to separate distillation of foreground and background. Recent research suggests it is preferable for the student model to reconstruct and learn expressive features from the teacher model in the first place instead of following the teacher for generating competitive representations.…”
Section: Introductionmentioning
confidence: 99%
“…55,56 Furthermore, KD is used for improving object detectors by selecting different valuable areas (e.g., foreground) to distill. [57][58][59] Inspired by these directions, we extend it to HOI recognition in videos, allowing knowledge transfer between global and local contextual views of interactions via KD.…”
Section: Knowledge Distillationmentioning
confidence: 99%