2020
DOI: 10.48550/arxiv.2012.10544
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Dataset Security for Machine Learning: Data Poisoning, Backdoor Attacks, and Defenses

Abstract: As machine learning systems grow in scale, so do their training data requirements, forcing practitioners to automate and outsource the curation of training data in order to achieve stateof-the-art performance. The absence of trustworthy human supervision over the data collection process exposes organizations to security vulnerabilities; training data can be manipulated to control and degrade the downstream behaviors of learned models. The goal of this work is to systematically categorize and discuss a wide ran… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
22
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(22 citation statements)
references
References 55 publications
0
22
0
Order By: Relevance
“…Solely altering the dataset, i.e., a Backdoor without Boosting is considered a Data Poisoning-based Backdoor. However, when using Boosting techniques, Backdoors are considered Model Poisoning-based Backdoors [74].…”
Section: A Targeting Integrity and Availabilitymentioning
confidence: 99%
See 2 more Smart Citations
“…Solely altering the dataset, i.e., a Backdoor without Boosting is considered a Data Poisoning-based Backdoor. However, when using Boosting techniques, Backdoors are considered Model Poisoning-based Backdoors [74].…”
Section: A Targeting Integrity and Availabilitymentioning
confidence: 99%
“…However, Server Cleaning requires a separated aggregator algorithm. For such purposes, the server might leverage different techniques based on the intrinsic properties of the updates [74].…”
Section: B Defending Integrity and Availabilitymentioning
confidence: 99%
See 1 more Smart Citation
“…Backdoor Attacks on Neural Networks. The key idea of backdoor attacks [12,21,23,53] is to inject hidden behaviors into a model, such that a test-time input stamped with a specific backdoor trigger (e.g. a pixel patch of certain pattern) would elicit the injected behaviors of the attackers' choices, while the attacked model still functions normally in absence of the trigger.…”
Section: Related Workmentioning
confidence: 99%
“…For years, one of the major goals of the AI security community is to securely and reliably produce and deploy deep learning models for real-world applications. To this end, data poisoning based backdoor attacks [12,21,53,72] on deep neural networks (DNNs) in the production stage (or training stage) and corresponding defenses [11,13,74] are extensively explored in recent years.…”
Section: Introductionmentioning
confidence: 99%