2017
DOI: 10.1147/jrd.2017.2709578
|View full text |Cite
|
Sign up to set email alerts
|

An effective algorithm for hyperparameter optimization of neural networks

Abstract: A major challenge in designing neural network (NN) systems is to determine the best structure and parameters for the network given the data for the machine learning problem at hand. Examples of parameters are the number of layers and nodes, the learning rates, and the dropout rates. Typically, these parameters are chosen based on heuristic rules and manually fine-tuned, which may be very time-consuming, because evaluating the performance of a single parametrization of the NN may require several hours. This pap… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
61
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 141 publications
(72 citation statements)
references
References 14 publications
0
61
0
1
Order By: Relevance
“…The performance of SA methods is highly sensitive to the chosen sequence of step sizes {α k } [Hutchison and Spall, 2013]. This mirrors the situation in gradient-based SA methods where the tuning of algorithmic parameters is an active area of research [Diaz et al, 2017, Ilievski et al, 2017, Balaprakash et al, 2018.…”
Section: Stochastic and Sample-average Approximationmentioning
confidence: 96%
“…The performance of SA methods is highly sensitive to the chosen sequence of step sizes {α k } [Hutchison and Spall, 2013]. This mirrors the situation in gradient-based SA methods where the tuning of algorithmic parameters is an active area of research [Diaz et al, 2017, Ilievski et al, 2017, Balaprakash et al, 2018.…”
Section: Stochastic and Sample-average Approximationmentioning
confidence: 96%
“…The capabilities of a neural network to make good predictions depends on its architecture and its parameters, it is an essential task to define a well structured network before implementing the model. Parameters which define the model architecture are known as hyperparameters and the process of assessing the best configuration for those parameters is called hyperparameter tuning (Diaz et al, 2017).…”
Section: Use Of Deep Learning Models For Classification Of Olive Oilmentioning
confidence: 99%
“…• Stochastic Approximation [21], hill climbing where hyperparameters are individually and sequentially changed • Evolutionary algorithms [20], which randomly start, select the best initial results (parents), and then generate multiple possible outcomes (children), and then repeat the process • Bayesian optimization (BO) [5] which treats the objective function as a random function and uses randomly determined hyperparameters to construct a distribution around the results • Other approaches which do not fit cleanly into these three groups, e.g. Radial Basis Functions [22], Hyberband [23], Nedler-Mead [24], and spectral approaches [25]. Beyond this work, further approaches include extensions of BO and combinations of methods.…”
Section: Ai Hyperparameter Determinationmentioning
confidence: 99%