2021
DOI: 10.1109/mis.2020.3028613
|View full text |Cite
|
Sign up to set email alerts
|

Hiding in the Crowd: Federated Data Augmentation for On-Device Learning

Abstract: To cope with the lack of on-device machine learning samples, this article presents a distributed data augmentation algorithm, coined federated data augmentation (FAug). In FAug, devices share a tiny fraction of their local data, i.e., seed samples, and collectively train a synthetic sample generator that can augment the local datasets of devices. To further improve FAug, we introduce a multi-hop based seed sample collection method and an oversampling technique that mixes up collected seed samples. Both approac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 19 publications
(18 citation statements)
references
References 8 publications
0
18
0
Order By: Relevance
“…Second, FL fundamentally requires ensuring privacy within each agent, which commonly implies but is not limited to maintaining data samples private. For instance, the information on possessed class labels for classification tasks should also be kept private [21]. However, because of their omniscient viewpoint, most existing personalized and decentralized learning schemes do not thoroughly preserve inter-device privacy.…”
Section: Introductionmentioning
confidence: 99%
“…Second, FL fundamentally requires ensuring privacy within each agent, which commonly implies but is not limited to maintaining data samples private. For instance, the information on possessed class labels for classification tasks should also be kept private [21]. However, because of their omniscient viewpoint, most existing personalized and decentralized learning schemes do not thoroughly preserve inter-device privacy.…”
Section: Introductionmentioning
confidence: 99%
“…In particular, MEC brings the high computing servers closer to the mobile users so that users with low computing and energy capability are able to offload their latency and computing-intensive tasks to the edge server [13]- [15]. Therefore, mobile users are able to offload a selected portion of local dataset to the edge server where a statistical model is trained by the edge server simultaneously with several mobile devices in hands [16], [17]. Even though FL is intended for the privacy preserving application, a portion of local dataset which are not privacysensitive can be offloaded to the MEC for further computation.…”
Section: Introductionmentioning
confidence: 99%
“…Inline with this idea, the works in [17] and [16] proposed the local data sharing mechanism for FL. In [16], the authors mitigated the non-i.i.d.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations