2023
DOI: 10.1109/tcns.2022.3232519
|View full text |Cite
|
Sign up to set email alerts
|

GTAdam: Gradient Tracking With Adaptive Momentum for Distributed Online Optimization

Abstract: This paper deals with a network of computing agents aiming to solve an online optimization problem in a distributed fashion, i.e., by means of local computation and communication, without any central coordinator. We propose the gradient tracking with adaptive momentum estimation (GTAdam) distributed algorithm, which combines a gradient tracking mechanism with first and second order momentum estimates of the gradient. The algorithm is analyzed in the online setting for strongly convex cost functions with Lipsch… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
6
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 39 publications
0
6
0
Order By: Relevance
“…A distributed adaptive gradient algorithm with bounded stepsizes was further studied in [31] to improve the generalization capacity. By introducing a GT estimator, a novel distributed adaptive algorithm was developed in the notable work [38], which is proved to achieve a linear convergence rate under the strongly-convex setting.…”
Section: Introductionmentioning
confidence: 99%
See 4 more Smart Citations
“…A distributed adaptive gradient algorithm with bounded stepsizes was further studied in [31] to improve the generalization capacity. By introducing a GT estimator, a novel distributed adaptive algorithm was developed in the notable work [38], which is proved to achieve a linear convergence rate under the strongly-convex setting.…”
Section: Introductionmentioning
confidence: 99%
“…The adaptive stepsizes are generated according to the historical gradients, enabling the algorithm to automatically coordinate the stepsizes among dimensions when the gradients are sparse. Inspired by [38], we utilize a GT estimator to aggregate the gradients over the network. Moreover, an clipping operator is used to mitigate the negative effects of extreme adaptive stepsizes.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations