2017
DOI: 10.1587/nolta.8.289
|View full text |Cite
|
Sign up to set email alerts
|

A novel quasi-Newton-based optimization for neural network training incorporating Nesterov's accelerated gradient

Abstract: This paper describes a novel quasi-Newton (QN) based accelerated technique for training of neural networks. Recently, Nesterov's accelerated gradient method has been utilized for the acceleration of the gradient-based training. In this paper the acceleration of the QN training algorithm is realized by the quadratic approximation of the error function incorporating the momentum term as Nesterov's method. It is shown that the proposed algorithm has a similar convergence property with the QN method. Neural networ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
51
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
2
2

Relationship

3
5

Authors

Journals

citations
Cited by 21 publications
(51 citation statements)
references
References 12 publications
0
51
0
Order By: Relevance
“…In this section, NAQ [9][13] and the adaptive momentum coefficient scheme [11] is introduced. The proposed AdaNAQ algorithm combines NAQ with the adaptive momentum coefficient scheme [14].…”
Section: Formulation Of Neural Network Trainingmentioning
confidence: 99%
See 2 more Smart Citations
“…In this section, NAQ [9][13] and the adaptive momentum coefficient scheme [11] is introduced. The proposed AdaNAQ algorithm combines NAQ with the adaptive momentum coefficient scheme [14].…”
Section: Formulation Of Neural Network Trainingmentioning
confidence: 99%
“…The former [5][6] [7] [8] are often used and improved because of the simplicity of their calculation. However, when applied to highly nonlinear problems, first-order methods still converge too slowly and the optimization error cannot be effectively reduced within a finite time [1] [9]. To deal with this problem, the quasi-Newton method (QN), which is one of the most efficient optimization methods with superlinear convergence has been widely utilized as a robust training algorithm [10].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Several modifications have been proposed to quasi-Newton to obtain stronger convergence. The Nesterov's Accelerated quasi-Newton method [15] gives faster convergence compared to the standard quasi-Newton methods. NAQ obtains faster convergence by quadratic approximation at w k + µv k and by incorporating the Nesterov's accelerated gradient ∇E(w k + µv k ) The derivation of NAQ is briefly introduced as follows:…”
Section: )mentioning
confidence: 99%
“…The NAQ [16] method achieves faster convergence compared to the standard QN by quadratic approximation of the objective function at w k + µv k and by incorporating the Nesterov's accelerated gradient ∇E(w k + µv k ) in its Hessian update. The update vector of NAQ is given as…”
mentioning
confidence: 99%