2021 IEEE 18th Annual Consumer Communications &Amp; Networking Conference (CCNC) 2021
DOI: 10.1109/ccnc49032.2021.9369498
|View full text |Cite
|
Sign up to set email alerts
|

Label Leakage from Gradients in Distributed Machine Learning

Abstract: Federated learning enables multiple users to build a joint model by sharing their model updates (gradients), while their raw data remains local on their devices. In contrast to the common belief that this provides privacy benefits, we here add to the very recent results on privacy risks when sharing gradients. Specifically, we propose Label Leakage from Gradients (LLG), a novel attack to extract the labels of the users' training data from their shared gradients. The attack exploits the direction and magnitude … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

2
7

Authors

Journals

citations
Cited by 15 publications
(6 citation statements)
references
References 30 publications
0
6
0
Order By: Relevance
“…The study proposed by [24] improved the weight matching algorithm for deep leakage from gradients to enhance the reconstruction performance. Similarly, the study in [25] also proposed the reconstruction of data while matching the gradient information. In addition, the study also uses auxiliary information in order to improve the reconstruction performance.…”
Section: Related Workmentioning
confidence: 99%
“…The study proposed by [24] improved the weight matching algorithm for deep leakage from gradients to enhance the reconstruction performance. Similarly, the study in [25] also proposed the reconstruction of data while matching the gradient information. In addition, the study also uses auxiliary information in order to improve the reconstruction performance.…”
Section: Related Workmentioning
confidence: 99%
“…However, their approach is limited to one-sample batch, which is uncommon in real-world applications of FL, where users typically have multiple data samples and train the model on these samples (a bunch of them at least) before sharing the gradients with the server. Wainakh et al [34], in a short paper (4 pages), introduced a basic idea to extend the attack of [38] for bigger batches, however, their work lacks formalization and thorough evaluation to substantially support the validity of the approach. Li et al [21] proposed also an analytical approach based on the observation that the gradient norms of a particular class are generally larger than the others.…”
Section: Label Extractionmentioning
confidence: 99%
“…If the identity is known while the models are aggregated, the adversary can infer something about the training data but not track it for a particular participant. In general, tracing requires ancillary information specific to the leak (Wainakh et al , 2021; McMahan et al , 2017; Zhu et al , 2019). For example, after inferring that photos of a certain person had started appearing in the training data, an adversary may have enough complementary and contextual information about the participants to guess which of them included these images in the training data (Yin et al , 2018; Blanchard et al , 2017).…”
Section: Threats and Defensesmentioning
confidence: 99%