2016
DOI: 10.48550/arxiv.1602.05629
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Communication-Efficient Learning of Deep Networks from Decentralized Data

Abstract: Modern mobile devices have access to a wealth of data suitable for learning models, which in turn can greatly improve the user experience on the device. For example, language models can improve speech recognition and text entry, and image models can automatically select good photos. However, this rich data is often privacy sensitive, large in quantity, or both, which may preclude logging to the data center and training there using conventional approaches. We advocate an alternative that leaves the training dat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
713
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 448 publications
(717 citation statements)
references
References 16 publications
4
713
0
Order By: Relevance
“…Baselines. We compare our proposed FedSpa with four baselines, including FedAvg [6], Sub-FedAvg [5], Ditto [3] and Local. We tune the hyper-parameters of the baselines to their best states.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Baselines. We compare our proposed FedSpa with four baselines, including FedAvg [6], Sub-FedAvg [5], Ditto [3] and Local. We tune the hyper-parameters of the baselines to their best states.…”
Section: Methodsmentioning
confidence: 99%
“…The learning rate is initialized with 0.1 and decayed with 0.998 after each communication round. We simulate 100 clients in total, and in each round 10 of them are picked to perform local training (the setting follows [6]). For all the methods except Ditto, local epochs are fixed to 5.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations