2020
DOI: 10.48550/arxiv.2003.00295
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Adaptive Federated Optimization

Abstract: Federated learning is a distributed machine learning paradigm in which a large number of clients coordinate with a central server to learn a model without sharing their own training data. Due to the heterogeneity of the client datasets, standard federated optimization methods such as Federated Averaging (FEDAVG) are often difficult to tune and exhibit unfavorable convergence behavior. In non-federated settings, adaptive optimization methods have had notable success in combating such issues. In this work, we pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

7
344
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 157 publications
(351 citation statements)
references
References 17 publications
7
344
0
Order By: Relevance
“…This section presents the experimental results of the proposed method, SPIDER in comparison to local adaptation, perFedAvg and Ditto with CIFAR100 dataset, another dataset explored by researchers for Federated learning [23]. All our experiments are based on a non-IID data distribution among FL clients.…”
Section: Experiments On Cifar100 Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…This section presents the experimental results of the proposed method, SPIDER in comparison to local adaptation, perFedAvg and Ditto with CIFAR100 dataset, another dataset explored by researchers for Federated learning [23]. All our experiments are based on a non-IID data distribution among FL clients.…”
Section: Experiments On Cifar100 Datasetmentioning
confidence: 99%
“…To address the data-heterogeneity challenge, variants of the standard FedAvg have been proposed to train a global model, including the FedProx [17], FedOPT [23], and FedNova [31]. In addition to training of a global model, frameworks that focus on training personalized models have also gained a lot of popularity.…”
Section: Introductionmentioning
confidence: 99%
“…And for Assumption. 3, besides the widely applied local gradient bounded variance in FL, we use the global bound σ g to quantify the data-heterogeneity due to the non-i.i.d distributed training dataset, which is also introduced in recent FL studies [19], [36]. Additionally, to illustrate the device-heterogeneity under the formulated systemheterogeneous FL in this paper, we make an extra assumption on the boundary of the approximated gradients from the proposed FedLGA algorithm as the following.…”
Section: Convergence Analysismentioning
confidence: 99%
“…In FedAvg, FL clients run multiple epochs of SGD before sending their locally computed gradients to the server, which updates the global model accordingly. Beyond FedAvg, other optimization mechanisms have been proposed to improve on convergence and efficiency aspects [45].…”
Section: Federated Learningmentioning
confidence: 99%