2020
DOI: 10.48550/arxiv.2007.13137
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fast-Convergent Federated Learning

Abstract: Federated learning has emerged recently as a promising solution for distributing machine learning tasks through modern networks of mobile devices. Recent studies have obtained lower bounds on the expected decrease in model loss that is achieved through each round of federated learning. However, convergence generally requires a large number of communication rounds, which induces delay in model training and is costly in terms of network resources. In this paper, we propose a fast-convergent federated learning al… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(8 citation statements)
references
References 21 publications
(43 reference statements)
0
8
0
Order By: Relevance
“…The local gradient, which is correlated with minimizing the local objective, may not align with the direction of approaching the optimal of the global objective. The correlation ∇ (w( )), ∇ (w( )) ∇ (w( )) ∇ (w( )) between the local gradient and the global gradient is a metric to measure their alignment 2 Similar assumption has made in FL context, for example in [13], [14], [17]. In [13], [17], the dissimilarity across local gradients is imposed by an upper bound to capture the impact of data heterogeneity on FL convergence, and an analogous definition named gradient divergence is also presented in [14].…”
Section: A Convergence Analysismentioning
confidence: 88%
See 4 more Smart Citations
“…The local gradient, which is correlated with minimizing the local objective, may not align with the direction of approaching the optimal of the global objective. The correlation ∇ (w( )), ∇ (w( )) ∇ (w( )) ∇ (w( )) between the local gradient and the global gradient is a metric to measure their alignment 2 Similar assumption has made in FL context, for example in [13], [14], [17]. In [13], [17], the dissimilarity across local gradients is imposed by an upper bound to capture the impact of data heterogeneity on FL convergence, and an analogous definition named gradient divergence is also presented in [14].…”
Section: A Convergence Analysismentioning
confidence: 88%
“…Generally, the FL algorithm adopts synchronous aggregation and selects a subset of nodes randomly to participate in each round randomly to avoid long-tailed waiting time due to the network uncertainty and straggler. To boost convergence and reduce the communication rounds, tuning the number of local updates [8], [13]- [15], and selecting appropriate nodes for FL training [12], [16], [17] are the usually adopted approaches.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations