GLOBECOM 2020 - 2020 IEEE Global Communications Conference 2020
DOI: 10.1109/globecom42002.2020.9322527
|View full text |Cite
|
Sign up to set email alerts
|

ClusterGrad: Adaptive Gradient Compression by Clustering in Federated Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(3 citation statements)
references
References 8 publications
0
3
0
Order By: Relevance
“…The DGC algorithm proposed in [24] discarded 99.9% of insignificant model updates for uploading on distributed machine learning workers, and different methods were designed to maintain the performance of the model. ClusterGrad [25] was based on clustering to adaptively filter and quantify important gradients according to the distribution of model updates. Li et al [26] took into account the heterogeneous resources on devices in FL.…”
Section: Communication Compression In Flmentioning
confidence: 99%
“…The DGC algorithm proposed in [24] discarded 99.9% of insignificant model updates for uploading on distributed machine learning workers, and different methods were designed to maintain the performance of the model. ClusterGrad [25] was based on clustering to adaptively filter and quantify important gradients according to the distribution of model updates. Li et al [26] took into account the heterogeneous resources on devices in FL.…”
Section: Communication Compression In Flmentioning
confidence: 99%
“…A first approach aims at reducing the energy spent for the communications between the clients and the central server. Gradient sparsification and gradient quantization techniques (see, e.g., [7,3,6]) have been recently proposed for that purpose. Since compression is a lossy process, the gains in terms of communication costs are usually achieved at the expense of a worse iteration complexity and there is not a good understanding of how these techniques impact the total energy consumption and whether they are worth applying in general.…”
Section: Introductionmentioning
confidence: 99%
“…Broadly speaking, the main processes of FL can be described as below: One of FL's mainstream scenarios is that the server locates in a data center with good Internet resources, and clients are edge devices (e.g., smartphones, IoTs, or wearable devices) (Konecný et al 2016a,b;McMahan et al 2017;Sattler et al 2020;Smith et al 2017;Zheng et al 2020) which are limited in low bandwidth and unstable network. Existing studies show that residential Internet connections tend to reach far higher download than upload speeds (Goga and Teixeira 2012;Konecný et al 2016a;Kairouz et al 2019;Cui et al 2020;Zhang et al 2021aZhang et al , 2022, thus there exist a serious bottleneck in FL when a huge number of clients send their local model to the server, especially when the uploaded models are the complicated and large-scale convolutional neural networks (CNNs).…”
Section: Introductionmentioning
confidence: 99%