2013
DOI: 10.1093/imanum/drs056
|View full text |Cite
|
Sign up to set email alerts
|

On spectral properties of steepest descent methods

Abstract: In recent years it has been made more and more clear that the critical issue in gradient methods is the choice of the step length, whereas using the gradient as search direction may lead to very effective algorithms, whose surprising behaviour has been only partially explained, mostly in terms of the spectrum of the Hessian matrix. On the other hand, the convergence of the classical Cauchy steepest descent (SD) method has been extensively analysed and related to the spectral properties of the Hessian matrix, b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
68
0
1

Year Published

2015
2015
2020
2020

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 67 publications
(70 citation statements)
references
References 26 publications
1
68
0
1
Order By: Relevance
“…This leads to simple results that reproduce properties described by Nocedal, Sartenaer and Zhu [11] and by De Asmundis et al [6,7]. Consider an application of the Cauchy algorithm as above, starting from x 0 with no null component, and define r = lim k→∞ y 2k , with y k = g k / g k .…”
Section: Properties Of the Limiting Pointssupporting
confidence: 56%
See 2 more Smart Citations
“…This leads to simple results that reproduce properties described by Nocedal, Sartenaer and Zhu [11] and by De Asmundis et al [6,7]. Consider an application of the Cauchy algorithm as above, starting from x 0 with no null component, and define r = lim k→∞ y 2k , with y k = g k / g k .…”
Section: Properties Of the Limiting Pointssupporting
confidence: 56%
“…The reason for proving these results, which are not original, is that the original Akaike and Forsythe papers, not being aimed exclusively at these results, are frequently not considered easy to read. We also prove simplified versions of results in [6,7,11], with a unified notation.…”
Section: The Cauchy Algorithmmentioning
confidence: 82%
See 1 more Smart Citation
“…In general, the relationship between the choice of steplength, in gradient-type methods, and the eigenvalues and eigenvectors of the underlying Hessian of the objective function is well-known, and for nonmonotone methods can be traced back to [21, pp. 117-118]; see also [17,31].…”
Section: Acceleration Strategiesmentioning
confidence: 99%
“…Usually, in iteration (2), λ (k) is adaptively computed to ensure the sufficient decrease of the objective function and, thus, the convergence of the whole scheme, while σ k is a 'free' parameter which can be chosen in order to improve the effectiveness of the algorithm (see e.g. [4,12,13,15]). In our analysis, we extend the convergence results about the gradient projection method (2)-(3) to the more general case where y (k) is defined as…”
mentioning
confidence: 99%