2019
DOI: 10.48550/arxiv.1905.02957
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

SAdam: A Variant of Adam for Strongly Convex Functions

Abstract: The Adam algorithm has become extremely popular for large-scale machine learning. Under convexity condition, it has been proved to enjoy a data-dependant O( √ T ) regret bound where T is the time horizon. However, whether strong convexity can be utilized to further improve the performance remains an open problem. In this paper, we give an affirmative answer by developing a variant of Adam (referred to as SAdam) which achieves a data-dependant O(log T ) regret bound for strongly convex functions. The essential … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 8 publications
0
5
0
Order By: Relevance
“…In the proposed DiffGrad optimization method, the steps up to the computation of bias-corrected 1-st order moment m t and bias-corrected 2-nd order moment v t are the same as those of Adam optimization [38]. The DiffGrad optimization method updates θ t+1 , using the following update rule:…”
Section: Adam-type Algorithmsmentioning
confidence: 99%
“…In the proposed DiffGrad optimization method, the steps up to the computation of bias-corrected 1-st order moment m t and bias-corrected 2-nd order moment v t are the same as those of Adam optimization [38]. The DiffGrad optimization method updates θ t+1 , using the following update rule:…”
Section: Adam-type Algorithmsmentioning
confidence: 99%
“…Reddi et al (2019) spots the issue of Adam convergence and provides a variant called AMSGrad while Zaheer et al (2018) argues the Adam only converges with large batch sizes. Subsequently, other variants of Adam are proposed in (Luo et al, 2019;Chen et al, 2019b;Huang et al, 2018;Wang et al, 2019b). Multiple lines of theoretical study on Adam are given in (Fang & Klabjan, 2019;Alacaoglu et al, 2020;Défossez et al, 2020)…”
Section: Related Workmentioning
confidence: 99%
“…Besides the aforementioned, other variants of Adam include NosAdam [40], Sadam [41], Adax [42]), AdaBound [15] and Yogi [43]. ACProp could be combined with other techniques such as SWATS [44], LookAhead [45] and norm regularization similar to AdamP [46].…”
Section: Related Workmentioning
confidence: 99%