2021
DOI: 10.48550/arxiv.2106.13973
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Benchmarking Differential Privacy and Federated Learning for BERT Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
7
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(14 citation statements)
references
References 0 publications
0
7
1
Order By: Relevance
“…Hoory et al (2021) achieve accuracy which is comparable to the non-private model, but additionally supplement the public pre-training data with additional domain-relevant material, while we use off-the-shelf pre-trained models. Basu et al (2021) observe significant drops in utility, compared to our parameter-efficient methods which do not. While Kerrigan et al (2020) consider public pre-training and private fine-tuning, their experiments are on much smaller architectures (i.e., feedforward networks with three hidden layers).…”
Section: Related Workcontrasting
confidence: 53%
See 1 more Smart Citation
“…Hoory et al (2021) achieve accuracy which is comparable to the non-private model, but additionally supplement the public pre-training data with additional domain-relevant material, while we use off-the-shelf pre-trained models. Basu et al (2021) observe significant drops in utility, compared to our parameter-efficient methods which do not. While Kerrigan et al (2020) consider public pre-training and private fine-tuning, their experiments are on much smaller architectures (i.e., feedforward networks with three hidden layers).…”
Section: Related Workcontrasting
confidence: 53%
“…Anil et al (2021) privately train BERT-Large from scratch, compared to our work which focuses on private fine-tuning. (Hoory et al, 2021;Basu et al, 2021) perform private full fine-tuning of BERT models. Hoory et al (2021) achieve accuracy which is comparable to the non-private model, but additionally supplement the public pre-training data with additional domain-relevant material, while we use off-the-shelf pre-trained models.…”
Section: Related Workmentioning
confidence: 99%
“…As an example, consider the problem of training a differentially private model [Dwork et al 2006] for a healthcare task. Training such a model "end-to-end" (i.e., without leveraging any pretraining) to a decent privacy-utility tradeoff currently requires vast amounts of privacy-sensitive data [McMahan et al 2018;Basu et al 2021]. In contrast, a foundation model pretrained on public data could potentially be adapted to the specific healthcare task with significantly less confidential data [Bommasani et al 2019;Tramèr and Boneh 2021].…”
Section: Opportunitiesmentioning
confidence: 99%
“…SEFL [19] is a FedNLP framework that achieves data privacy without any trusted entities. [10] studies how FedNLP can orchestrate with differential privacy. None of above work addresses the high training cost of FedNLP.…”
Section: Related Workmentioning
confidence: 99%