2019
DOI: 10.48550/arxiv.1912.02771
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Label-Consistent Backdoor Attacks

Abstract: Deep neural networks have been demonstrated to be vulnerable to backdoor attacks. Specifically, by injecting a small number of maliciously constructed inputs into the training set, an adversary is able to plant a backdoor into the trained model. This backdoor can then be activated during inference by a backdoor trigger to fully control the model's behavior. While such attacks are very effective, they crucially rely on the adversary injecting arbitrary inputs that are-often blatantly-mislabeled. Such samples wo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
123
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 57 publications
(127 citation statements)
references
References 15 publications
4
123
0
Order By: Relevance
“…One of the trends in backdoor attacks is to be more concealed, where this concealment is reflected in two aspects. On the one hand, the attacker hopes that the constructed trigger can evade human perception, so some researchers have proposed the invisible attacks [38], [39] and the labelconsistent attack [29]. On the other hand, the attack method needs to be machine imperceptible, that is, to escape various backdoor defenses.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…One of the trends in backdoor attacks is to be more concealed, where this concealment is reflected in two aspects. On the one hand, the attacker hopes that the constructed trigger can evade human perception, so some researchers have proposed the invisible attacks [38], [39] and the labelconsistent attack [29]. On the other hand, the attack method needs to be machine imperceptible, that is, to escape various backdoor defenses.…”
Section: Discussionmentioning
confidence: 99%
“…Two image datasets, CIFAR-10 [27] and CelebA [28], which are often used in the research of backdoor learning [25], [29], [30], are selected. For CelebA, following the configuration in some previous studies [25], [30], we use the three most balanced attributes, i.e., "Heavy Makeup", "Mouth Slightly Open", and "Smiling", to create eight categories for the image classification task.…”
Section: Datasetsmentioning
confidence: 99%
“…While the stability attacks considered in this work may be reminiscent of backdoor attacks [11], we note that they share several key differences. First, stability attacks aim to compromise adversarial training with welldefined -robustness, while backdoor attacks mainly focus on embedding exploits (that can be invoked by pre-specified triggers) into naturally trained models [23,50,67]. Second, stability attacks only perturb the inputs slightly, while many works on backdoor attacks require mislabeling [26,39,46,38].…”
Section: Related Workmentioning
confidence: 99%
“…• Clean-label [43], [44], [45]: the intuition behind cleanlabel attacks is that they do not require control over the labeling function. The poisoned dataset seems to be correct to the eyes of an expert label-verifier identity.…”
Section: A Targeting Integrity and Availabilitymentioning
confidence: 99%