2021
DOI: 10.48550/arxiv.2105.09369
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

User-Level Label Leakage from Gradients in Federated Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
14
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 10 publications
(15 citation statements)
references
References 0 publications
1
14
0
Order By: Relevance
“…In this scenario, the user does not have to send updated batch norm statistics to the server. We further assume, for simplicity, that label information is contained in the update as metadata, but note that novel label recovery algorithms as discussed in Wainakh et al (2021) are highly successful, even at large aggregation sizes.…”
Section: B Technical Detailsmentioning
confidence: 99%
“…In this scenario, the user does not have to send updated batch norm statistics to the server. We further assume, for simplicity, that label information is contained in the update as metadata, but note that novel label recovery algorithms as discussed in Wainakh et al (2021) are highly successful, even at large aggregation sizes.…”
Section: B Technical Detailsmentioning
confidence: 99%
“…Many works have investigated what an attacker can infer from inspecting the intermediate gradients in FL settings or from inspecting multiple model snapshots during training [22,23,24,25,26]. These attacks focus on inferring training points, their labels, or related properties.…”
Section: E Further Related Workmentioning
confidence: 99%
“…Depending on the model architecture, this frequency estimation can either be triggered by analyzing the bias gradient of the decoder (last linear) layer, or the norm of the embedding matrix. In both cases, the update magnitude is approximately proportional to the frequency of word usage, allowing for a greedy estimation by adapting the strategy of Wainakh et al (2021) which was originally proposed to extract label information in classification tasks.…”
Section: Getting Tokensmentioning
confidence: 99%
“…Algorithm 2 and Algorithm 3 detail the token recovery for transformers with decoder bias and for transformer with a tied embedding. These roughly follow the principles of greedy label recovery strategy proposed in Wainakh et al (2021) and we reproduce them here for completeness, incorporating additional considerations necessary for token retrieval.…”
Section: B Algorithm Detailsmentioning
confidence: 99%