2022
DOI: 10.48550/arxiv.2206.00240
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Privacy for Free: How does Dataset Condensation Help Privacy?

Abstract: To prevent unintentional data leakage, research community has resorted to data generators that can produce differentially private data for model training. However, for the sake of the data privacy, existing solutions suffer from either expensive training cost or poor generalization performance. Therefore, we raise the question whether training efficiency and privacy can be achieved simultaneously. In this work, we for the first time identify that dataset condensation (DC) which is originally designed for impro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 23 publications
0
5
0
Order By: Relevance
“…We also note the approach seems similar to the gradient inversion attack (Zhu et al, 2019) but we consider averaged gradients w.r.t. local virtual data, and the method potentially defenses inference attack better (Appendix Figure 13), which is also implied by (Xiong et al, 2022;Dong et al, 2022). Privacy preservation can be further improved by employing differential privacy (Abadi et al, 2016) in gradient sharing, but this is not the main focus of our work.…”
Section: Global Data Distillationmentioning
confidence: 74%
See 2 more Smart Citations
“…We also note the approach seems similar to the gradient inversion attack (Zhu et al, 2019) but we consider averaged gradients w.r.t. local virtual data, and the method potentially defenses inference attack better (Appendix Figure 13), which is also implied by (Xiong et al, 2022;Dong et al, 2022). Privacy preservation can be further improved by employing differential privacy (Abadi et al, 2016) in gradient sharing, but this is not the main focus of our work.…”
Section: Global Data Distillationmentioning
confidence: 74%
“…It brings in significantly improved the distillation efficiency. Moreover, recent studies have justified that data distillation also preserves privacy (Dong et al, 2022;Carlini et al, 2022b), which is critical in federated learning. Other modern data distillation strategies can be found here (Sachdeva & McAuley, 2023).…”
Section: Dataset Distillationmentioning
confidence: 99%
See 1 more Smart Citation
“…Dataset condensation aims to condense large datasets to smaller ones while preserving the information to train models. It can benefit various applications including continual learning [53,54,56], efficient neural architecture search [53,54,56], federated learning [17,41,59] and privacy-preserving ML [11,27]. Data Distillation (DD) [47] pioneered this topic by maximizing the accuracy of models trained by the condensed set with metalearning techniques [33].…”
Section: Related Workmentioning
confidence: 99%
“…However, the sheer amount of data introduces significant obstacles for storage, transmission, and data pre-processing. Besides, publishing raw data inevitably brings about privacy or copyright issue in practice [44,10]. To alleviate these problems, Wang et al [52] pioneer the research of dataset distillation (DD), to distill a large dataset into a synthetic one with only a limited number of samples, so that the training efforts with the distilled dataset for downstream models on the original dataset can be largely reduced, which facilitates a series of applications like continual learning [41,40,54,31] and black-box optimization [7].…”
Section: Introductionmentioning
confidence: 99%