2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01201
|View full text |Cite
|
Sign up to set email alerts
|

MiLeNAS: Efficient Neural Architecture Search via Mixed-Level Reformulation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
76
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 108 publications
(76 citation statements)
references
References 12 publications
0
76
0
Order By: Relevance
“…The approaches to personalization discussed in Section 3.3 still share the same model architecture among all clients. The recent progress in NAS [230,387,175,388,60,375,313,488,175,323] provides a potential way to address these drawbacks. There are three major methods for NAS, which utilize evolutionary algorithms, reinforcement learning, or gradient descent to search for optimal architectures for a specific task on a specific dataset.…”
Section: Neural Architecture Designmentioning
confidence: 99%
“…The approaches to personalization discussed in Section 3.3 still share the same model architecture among all clients. The recent progress in NAS [230,387,175,388,60,375,313,488,175,323] provides a potential way to address these drawbacks. There are three major methods for NAS, which utilize evolutionary algorithms, reinforcement learning, or gradient descent to search for optimal architectures for a specific task on a specific dataset.…”
Section: Neural Architecture Designmentioning
confidence: 99%
“…Most recently GD-based NAS methods are formulated as bilevel optimization problems, However, He et al [116] observe that bilevel optimization in the current methods is solved based on a heuristic. For instance, solution of the problem needs to get an approximation of the second-order methods [109,110].…”
Section: Nas Based On Gdmentioning
confidence: 99%
“…For instance, solution of the problem needs to get an approximation of the second-order methods [109,110]. He et al [116] demonstrate that the approximation has a superposition influence mainly because it is based on a one-step approximation of the network weights. As a result, gradient errors may cause the algorithm to fail to converge to a (locally) optimal solution.…”
Section: Nas Based On Gdmentioning
confidence: 99%
“…However, single-level optimization will overfit the architecture A and result in performance degradation during the validation process. Then, a mixed-level optimization [68] is proposed to deal with this problem and reduce gradient errors; that is, min w,A…”
Section: Stability and Convergence Analysismentioning
confidence: 99%