Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence 2022
DOI: 10.24963/ijcai.2022/399
|View full text |Cite
|
Sign up to set email alerts
|

Heterogeneous Ensemble Knowledge Transfer for Training Large Models in Federated Learning

Abstract: We propose a new method for unsupervised generative continual learning through realignment of Variational Autoencoder's latent space. Deep generative models suffer from catastrophic forgetting in the same way as other neural structures. Recent generative continual learning works approach this problem and try to learn from new data without forgetting previous knowledge. However, those methods usually focus on artificial scenarios where examples share almost no similarity between subsequent portions of data - a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
32
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 62 publications
(32 citation statements)
references
References 0 publications
0
32
0
Order By: Relevance
“…Knowledge Type Require Public Dataset STU-KD [44] Logits Yes DS-FL [13] FD [14] CFD [31] FedGEMS [6] Fed-ET [7] FedKEMF [41] [30] Mix2FLD [28] Ahn, et al [1,2] Def-KT [19] CMFD [34] pFedSD [17] No FedZKT [43] NRFL [25] FedGKT [10] Features+Logits No FedICT [39] FedD3 [32] Distilled Datasets No cation efficiency improvement compared with prior works. On top of this, Song [32] distils local datasets on devices and uploads the distilled dataset to the edge server, requiring only one-shot communication during the entire training process.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Knowledge Type Require Public Dataset STU-KD [44] Logits Yes DS-FL [13] FD [14] CFD [31] FedGEMS [6] Fed-ET [7] FedKEMF [41] [30] Mix2FLD [28] Ahn, et al [1,2] Def-KT [19] CMFD [34] pFedSD [17] No FedZKT [43] NRFL [25] FedGKT [10] Features+Logits No FedICT [39] FedD3 [32] Distilled Datasets No cation efficiency improvement compared with prior works. On top of this, Song [32] distils local datasets on devices and uploads the distilled dataset to the edge server, requiring only one-shot communication during the entire training process.…”
Section: Methodsmentioning
confidence: 99%
“…To reduce on-device computation overhead in FEL, a series of approaches [10,6,7] are proposed to enable devices training much smaller models than the edge server, leveraging KD as an exchange protocol across model representations. Specifically, He [10] and Cheng [6] establish alternating minimization FEL frameworks to transfer knowledge from compact on-device models to the large edge model via KD, after which the on-device models are optimized based on the knowledge transferred back from the edge.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations