2023
DOI: 10.1007/978-3-031-28996-5_2
|View full text |Cite
|
Sign up to set email alerts
|

Federated Learning with GAN-Based Data Synthesis for Non-IID Clients

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(11 citation statements)
references
References 15 publications
0
11
0
Order By: Relevance
“…More Discussions for Our Assumptions. Assumptions 1, 2, 3 are standard in nonconvex optimization (Reddi et al, 2016) and FL optimization (Reddi et al, 2021;Li et al, 2020b;2021a)…”
Section: A1 Additional Discussion and Theoretical Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…More Discussions for Our Assumptions. Assumptions 1, 2, 3 are standard in nonconvex optimization (Reddi et al, 2016) and FL optimization (Reddi et al, 2021;Li et al, 2020b;2021a)…”
Section: A1 Additional Discussion and Theoretical Resultsmentioning
confidence: 99%
“…This lack of research is likely because: ensemble distillation loss (e.g., R KD in (Global Obj)) is calculated upon local models' averaged logits (ensemble of teachers), so the loss w.r.t each local model (individual teacher) is not directly reflected. However, in standard federated optimization, the "global objective" can usually be "decomposed" into "local objectives" (Li et al, 2020b;a;T Dinh et al, 2020;Fallah et al, 2020), and it is easier to first measure the optimization error on each local objective (i.e., individual teacher) and then derive its implications on the global objective (i.e., ensemble of teachers). Therefore, to make KD loss mathematically tractable for convergence analysis, we focus on its upper bound, i.e., the average of individual distillation loss calculated upon each teacher model.…”
Section: Challenges Of Ensemble Distillation and A Relaxedmentioning
confidence: 99%
See 2 more Smart Citations
“…3 shows that equal averaging over all participants, as provided in FedAvg, achieves the poorest performance as the environment is significantly non-i.i.d. [27]. Meanwhile, FedAvg+ compensates for the overfitting loss from FedAvg.…”
Section: Personalized Decentralized Learning With Knowledge Distillationmentioning
confidence: 99%