2022
DOI: 10.3390/app12189389
|View full text |Cite
|
Sign up to set email alerts
|

AdaCB: An Adaptive Gradient Method with Convergence Range Bound of Learning Rate

Abstract: Adaptive gradient descent methods such as Adam, RMSprop, and AdaGrad achieve great success in training deep learning models. These methods adaptively change the learning rates, resulting in a faster convergence speed. Recent studies have shown their problems include extreme learning rates, non-convergence issues, as well as poor generalization. Some enhanced variants have been proposed, such as AMSGrad, and AdaBound. However, the performances of these alternatives are controversial and some drawbacks still occ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 30 publications
0
1
0
Order By: Relevance
“…To achieve a fast training process and attain convergence, some researchers have proposed several adaptive optimization algorithms ADAGRAD [28], RMSPROP [29], AdaBond, and AMSBond [30]. The element‐level scaling terms in the learning rate were utilised, but those methods demonstrated poor generalisation ability and the model cannot converge due to the unstable and extreme learning rate.…”
Section: Network Modelmentioning
confidence: 99%
“…To achieve a fast training process and attain convergence, some researchers have proposed several adaptive optimization algorithms ADAGRAD [28], RMSPROP [29], AdaBond, and AMSBond [30]. The element‐level scaling terms in the learning rate were utilised, but those methods demonstrated poor generalisation ability and the model cannot converge due to the unstable and extreme learning rate.…”
Section: Network Modelmentioning
confidence: 99%