2015
DOI: 10.1007/978-3-319-20424-6_12
|View full text |Cite
|
Sign up to set email alerts
|

A Framework for Selecting Deep Learning Hyper-parameters

Abstract: Abstract. Recent research has found that deep learning architectures show significant improvements over traditional shallow algorithms when mining high dimensional datasets. When the choice of algorithm employed, hyper-parameter setting, number of hidden layers and nodes within a layer are combined, the identification of an optimal configuration can be a lengthy process. Our work provides a framework for building deep learning architectures via a stepwise approach, together with an evaluation methodology to qu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 17 publications
0
3
0
Order By: Relevance
“…ANN algorithm is also known to produce good accuracy in energy load prediction (K. Li et al, 2018). Multi-layer Perceptron (MLP) is a function of a deep neural network that utilizes a feed forward propagation process with one hidden layer where latent and abstract features are learned (Donoghue and Roantree, 2015). In research by Khantach et al, Multi-layer Perceptron ANN produced the most accurate result between support vector machine (SVM), Gaussian process and radial basis function (RBF) with a Mean Absolute Percentage Error (MAPE) of 0.96 (Khantach et al, 2019).…”
Section: Literature Reviewmentioning
confidence: 99%
“…ANN algorithm is also known to produce good accuracy in energy load prediction (K. Li et al, 2018). Multi-layer Perceptron (MLP) is a function of a deep neural network that utilizes a feed forward propagation process with one hidden layer where latent and abstract features are learned (Donoghue and Roantree, 2015). In research by Khantach et al, Multi-layer Perceptron ANN produced the most accurate result between support vector machine (SVM), Gaussian process and radial basis function (RBF) with a Mean Absolute Percentage Error (MAPE) of 0.96 (Khantach et al, 2019).…”
Section: Literature Reviewmentioning
confidence: 99%
“…The discarded works did not comply with the quality criteria for a series of different reasons. Some papers, for example, were more focused on hyperparameter optimization (out of the scope of this review) than in selecting an appropriate algorithm given the context of application [53][54][55][56][57][58].…”
Section: Clarifications On the Excluded Recordsmentioning
confidence: 99%
“…Unfortunately, despite the fact that deep learning algorithms have been around for a long time, there are no wellestablished procedures for hyper-parameters tuning, such as back-propagation for a model training [5]. Instead, a set of custom techniques, such as grid, random and heuristic search [6,7], have been developed and used by most of DL systems designers.…”
Section: Introductionmentioning
confidence: 99%