1999
DOI: 10.1109/72.750573
|View full text |Cite
|
Sign up to set email alerts
|

A method to determine the required number of neural-network training repetitions

Abstract: Conventional neural-network training algorithms often get stuck in local minima. To find the global optimum, training is conventionally repeated with ten, or so, random starting values for the weights. Here we develop an analytical procedure to determine how many times a neural network needs to be trained, with random starting weights, to ensure that the best of those is within a desirable lower percentile of all possible trainings, with a certain level of confidence. The theoretical developments are validated… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
46
0

Year Published

2004
2004
2016
2016

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 93 publications
(46 citation statements)
references
References 13 publications
0
46
0
Order By: Relevance
“…2, as the median performance for any given river is relatively constant with respect to the number of hidden neurons. However, the performance varies considerably from one replicate to the other, exemplifying the risk of stand-alone models, as mentioned by Iyer and Rhinehart (1999). If exception is made of models with isolated extremely low performance, this variability tends to increase with the number of hidden neurons for most rivers in Fig.…”
Section: Stand-alone Modelsmentioning
confidence: 99%
“…2, as the median performance for any given river is relatively constant with respect to the number of hidden neurons. However, the performance varies considerably from one replicate to the other, exemplifying the risk of stand-alone models, as mentioned by Iyer and Rhinehart (1999). If exception is made of models with isolated extremely low performance, this variability tends to increase with the number of hidden neurons for most rivers in Fig.…”
Section: Stand-alone Modelsmentioning
confidence: 99%
“…Sixty percent of the data set was selected for the training set and the remainder only for the test set (the classification evaluation) that is never used for training the networks. The network training was repeated 15 times, and then the network producing the best accuracy was selected; this implies that the retained model resides within the best 14% of the distribution of all possible models (99% confidence) [35].…”
Section: Discussionmentioning
confidence: 99%
“…In the first case, training should just be restarted with random weights, while in the second case, the number of hidden nodes should be increased -otherwise training may not converge. Taking a probabilistic approach to this problem, one can estimate the number of training attempts that should be made before one establishes the inability of the training procedure to converge for a given network state by computing [8]:…”
Section: Adaptive Architecturementioning
confidence: 99%