Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence 2019
DOI: 10.24963/ijcai.2019/488
|View full text |Cite
|
Sign up to set email alerts
|

Heavy-ball Algorithms Always Escape Saddle Points

Abstract: Nonconvex optimization algorithms with random initialization have attracted increasing attention recently. It has been showed that many first-order methods always avoid saddle points with random starting points. In this paper, we answer a question: can the nonconvex heavy-ball algorithms with random initialization avoid saddle points? The answer is yes! Direct using the existing proof technique for the heavy-ball algorithms is hard due to that each iteration of the heavy-ball algorithm consists of current and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

3
6

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 6 publications
0
7
0
Order By: Relevance
“…1.2.0.1 Heavy ball: The convergence of deterministic HB, i.e., HB with exact gradient, has been thoroughly studied by [11], [14], [15], [16], [17] in both convex and nonconvex cases. An interesting finding is that HB can escape saddle points in nonconvex optimization by using a larger learning rate than GD [18]. HB momentum has also been successfully integrated into SGD to improve training DNNs.…”
Section: Additional Related Workmentioning
confidence: 99%
“…1.2.0.1 Heavy ball: The convergence of deterministic HB, i.e., HB with exact gradient, has been thoroughly studied by [11], [14], [15], [16], [17] in both convex and nonconvex cases. An interesting finding is that HB can escape saddle points in nonconvex optimization by using a larger learning rate than GD [18]. HB momentum has also been successfully integrated into SGD to improve training DNNs.…”
Section: Additional Related Workmentioning
confidence: 99%
“…In the nonconvex community, the inertial technique [17] (also called as heavy-ball or momentum) is widely used and proved to be algorithmically efficient [18,19,20,21]. Besides acceleration and good practical performance for nonconvex problems, the advantage of inertial technique is illustrated by weaker conditions avoiding saddle points [22]. The procedure of inertial method is quite simple, it uses linear combination of current and last point for next iteration.…”
Section: Inertial Methodsmentioning
confidence: 99%
“…optimal oracle complexity result for (1.1) when F is smooth. The work [58] studies how heavy-ball technique can help SGM escape saddle points. Distributed/parallel stochastic methods with delayed (sub)gradient information.…”
Section: Methodsmentioning
confidence: 99%