2021
DOI: 10.1609/aaai.v35i4.16387
|View full text |Cite
|
Sign up to set email alerts
|

Teacher Guided Neural Architecture Search for Face Recognition

Abstract: Knowledge distillation is an effective tool to compress large pre-trained convolutional neural networks (CNNs) or their ensembles into models applicable to mobile and embedded devices. However, with expected flops or latency, existing methods are hand-crafted heuristics. They propose to pre-define the target student network for knowledge distillation, which may be sub-optimal because it requires much effort to explore a powerful student from the large design space. In this paper, we develop a novel teacher gui… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 46 publications
0
2
0
Order By: Relevance
“…With the prior knowledge about typical properties of architectures, NAS approaches commonly define the searching space as a large set of operations (e.g., convolution, fully-connected, and pooling). Each possible architecture in the searching space is evaluated by a certain evaluation strategy [32], [33] and the searching process is controlled by certain searching algorithms, such as reinforcement learning [33], [35], [36], evolutionary search [37], differentiable search [38], or other learning algorithms [34], [39], [40], [41]. NAS commonly defines a searching space at first and then uses a certain policy to generate a sequence of actions in the searching space to specify the architecture.…”
Section: B a Brief Overview Of The Proposed Approachmentioning
confidence: 99%
“…With the prior knowledge about typical properties of architectures, NAS approaches commonly define the searching space as a large set of operations (e.g., convolution, fully-connected, and pooling). Each possible architecture in the searching space is evaluated by a certain evaluation strategy [32], [33] and the searching process is controlled by certain searching algorithms, such as reinforcement learning [33], [35], [36], evolutionary search [37], differentiable search [38], or other learning algorithms [34], [39], [40], [41]. NAS commonly defines a searching space at first and then uses a certain policy to generate a sequence of actions in the searching space to specify the architecture.…”
Section: B a Brief Overview Of The Proposed Approachmentioning
confidence: 99%
“…To reduce the NAS search time, all mentioned NAS algorithms proposed to learn from small training datasets such as CIFAR-10 [20] and then utilized the discovered architecture to train on larger datasets such as ImageNet [21]. This advancement in NAS solutions has only recently captured the attention of biometric recognition solutions [22], [23], however, with no deployments towards lightweight or embedded architectures.…”
Section: Introductionmentioning
confidence: 99%