2021
DOI: 10.48550/arxiv.2104.13417
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Towards Fair Federated Learning with Zero-Shot Data Augmentation

Abstract: Federated learning has emerged as an important distributed learning paradigm, where a server aggregates a global model from many client-trained models, while having no access to the client data. Although it is recognized that statistical heterogeneity of the client local data yields slower global model convergence, it is less commonly recognized that it also yields a biased federated global model with a high variance of accuracy across clients. In this work, we aim to provide federated learning schemes with im… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 30 publications
0
2
0
Order By: Relevance
“…The key motivation behind data sharing is that a client cannot acquire samples from other clients during local training, thus the learned local model under-represents certain patterns or samples from the absent classes. The common practices are to share a public dataset [6], synthesized data [16,17] or a condensed version of the training samples [18] to supplement training on the clients or on the server. This line of works may violate the privacy rule of federated learning since they all consider sharing raw input data of the model, either real data or artificial data.…”
Section: Aggregationmentioning
confidence: 99%
See 1 more Smart Citation
“…The key motivation behind data sharing is that a client cannot acquire samples from other clients during local training, thus the learned local model under-represents certain patterns or samples from the absent classes. The common practices are to share a public dataset [6], synthesized data [16,17] or a condensed version of the training samples [18] to supplement training on the clients or on the server. This line of works may violate the privacy rule of federated learning since they all consider sharing raw input data of the model, either real data or artificial data.…”
Section: Aggregationmentioning
confidence: 99%
“…There have been a plethora of works exploring promising solutions to federated learning on non-IID data. They can be roughly divided into four categories: 1) client drift mitigation [5,8,9,10], which modifies the local objectives of the clients, so that the local model is consistent with the global model to a certain degree; 2) aggregation scheme [11,12,13,14,15], which improves the model fusion mechanism at the server; 3) data sharing [6,16,17,18], which introduces public datasets or synthesized data to help construct a more balanced data distribution on the client or on the server; 4) personalized federated learning [19,20,21,22], which aims to train personalized models for individual clients rather than a shared global model. However, as suggested by [7], existing algorithms are still unable to achieve good performance on image datasets with deep learning models, and could be no better than vanilla FedAvg [2].…”
Section: Introductionmentioning
confidence: 99%