2018
DOI: 10.1007/978-3-030-01246-5_2
|View full text |Cite
|
Sign up to set email alerts
|

Progressive Neural Architecture Search

Abstract: We propose a new method for learning the structure of convolutional neural networks (CNNs) that is more efficient than recent state-of-the-art methods based on reinforcement learning and evolutionary algorithms. Our approach uses a sequential model-based optimization (SMBO) strategy, in which we search for structures in order of increasing complexity, while simultaneously learning a surrogate model to guide the search through structure space. Direct comparison under the same search space shows that our method … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
1,313
0
4

Year Published

2018
2018
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,592 publications
(1,320 citation statements)
references
References 19 publications
3
1,313
0
4
Order By: Relevance
“…architecture depends on the difficulty and size of the dataset at hand. While these findings may encourage an automated neural architecture search, such an approach is hindered by the limited computational resources [19], [20], [21], [22], [23]. Alternatively, we propose an ensemble architecture, which combines U-Nets of varying depths into one unified structure.…”
Section: Table Imentioning
confidence: 99%
“…architecture depends on the difficulty and size of the dataset at hand. While these findings may encourage an automated neural architecture search, such an approach is hindered by the limited computational resources [19], [20], [21], [22], [23]. Alternatively, we propose an ensemble architecture, which combines U-Nets of varying depths into one unified structure.…”
Section: Table Imentioning
confidence: 99%
“…The progressive neural architecture search (PNAS) investigated the use of the Bayesian optimization strategy SMBO to make the search for CNN architectures more efficient by exploring simpler cells before determining whether to search more complex cells [71]. Similarly, NASBOT defines a distance function for generated architectures, which is used for constructing a kernel to use Gaussian processes for BO [72].…”
Section: Neural Architecture Searchmentioning
confidence: 99%
“…This is often achieved by jointly training a fully connected layer for dimension control and reordering for each modality, together with the scalar weights for fusion. A recent study [126] employs neural architecture search with progressive exploration [127]- [129] to find suitable settings for a number of fusion functions. Each fusion function is configured by which layers to fuse and whether to use concatenation or weighted sum as the fusion operation.…”
Section: A Simple Operation-based Fusionmentioning
confidence: 99%