2021
DOI: 10.48550/arxiv.2110.01663
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Global Convergence and Stability of Stochastic Gradient Descent

Abstract: In machine learning, stochastic gradient descent (SGD) is widely deployed to train models using highly non-convex objectives with equally complex noise models. Unfortunately, SGD theory often makes restrictive assumptions that fail to capture the non-convexity of real problems, and almost entirely ignore the complex noise models that exist in practice. In this work, we make substantial progress on this shortcoming. First, we establish that SGD's iterates will either globally converge to a stationary point or d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(5 citation statements)
references
References 20 publications
0
5
0
Order By: Relevance
“…Our assumption is equivalent to the one in [11]). But it is weaker than those assumptions made in ( [4,34]).…”
Section: Assumptions For Sgdmentioning
confidence: 78%
See 4 more Smart Citations
“…Our assumption is equivalent to the one in [11]). But it is weaker than those assumptions made in ( [4,34]).…”
Section: Assumptions For Sgdmentioning
confidence: 78%
“…There are several interesting remaining questions. First, whether SGD avoids saddle points has drawn much recent attention ( [13,20,23,31,34]). Studying this question under very mild assumptions would be interesting.…”
Section: Summary and Discussionmentioning
confidence: 99%
See 3 more Smart Citations