1999
DOI: 10.1016/s0893-6080(99)00072-6
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating neural network training using weight extrapolations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
55
0
2

Year Published

2004
2004
2013
2013

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 123 publications
(57 citation statements)
references
References 16 publications
0
55
0
2
Order By: Relevance
“…Much of the existing analysis in the neural network literature (Haykin, 1999;Phansalkar & Sastry, 1994;Qian, 1999;Torii & Hagan 2002) is restricted to the case of constant l and m: There is also a large literature on the time-varying case, generally referred to as dynamic or adaptive choice of the learning rate and momentum factors (see Kamarthi and Pittner (1999) and references therein), but, to the best of our knowledge, the observations made in this paper are new. We now present the CG method from a control viewpoint, which is the inspiration for the results obtained here.…”
Section: Steepest Descent Plus Momentum Equals Frozen Conjugate Gradientmentioning
confidence: 96%
See 3 more Smart Citations
“…Much of the existing analysis in the neural network literature (Haykin, 1999;Phansalkar & Sastry, 1994;Qian, 1999;Torii & Hagan 2002) is restricted to the case of constant l and m: There is also a large literature on the time-varying case, generally referred to as dynamic or adaptive choice of the learning rate and momentum factors (see Kamarthi and Pittner (1999) and references therein), but, to the best of our knowledge, the observations made in this paper are new. We now present the CG method from a control viewpoint, which is the inspiration for the results obtained here.…”
Section: Steepest Descent Plus Momentum Equals Frozen Conjugate Gradientmentioning
confidence: 96%
“…For a discussion of some of these practical computational issues, we refer the reader to Kamarthi and Pittner (1999) and Yu and Chen (1997) (and references therein).…”
Section: Practical Implications For Tuning Of Parametersmentioning
confidence: 99%
See 2 more Smart Citations
“…The most popular learning algorithm used in prediction purposes is the back propagation algorithm (BPA) or the generalized delta rule. The BPA is a supervised learning algorithm that aims at reducing overall system error to a minimum [1,9]. This algorithm has made multilayer neural networks suitable for various prediction problems.…”
Section: Artificial Neural Network With Backpropagation Learning -An mentioning
confidence: 99%