2021 IEEE International Conference on Image Processing (ICIP) 2021
DOI: 10.1109/icip42928.2021.9506313
|View full text |Cite
|
Sign up to set email alerts
|

Simtrojan: Stealthy Backdoor Attack

Abstract: Recent researches indicate deep learning models are vulnerable to adversarial attacks. Backdoor attack, also called trojan attack, is a variant of adversarial attacks. An malicious attacker can inject backdoor to models in training phase. As a result, the backdoor model performs normally on clean samples and can be triggered by a backdoor pattern to recognize backdoor samples as a wrong target label specified by the attacker. However, the vanilla backdoor attack method causes a measurable difference between cl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(8 citation statements)
references
References 9 publications
0
8
0
Order By: Relevance
“…Ideally, we hope a perfect adaptive attack can make the poison and clean samples completely indistinguishable. This has been achieved under stronger threat model when the the training process is also controlled [27,35,7,25,5,37]. In this paper, we take a step further to this goal under poisoning-only threat model.…”
Section: Discussionmentioning
confidence: 94%
See 2 more Smart Citations
“…Ideally, we hope a perfect adaptive attack can make the poison and clean samples completely indistinguishable. This has been achieved under stronger threat model when the the training process is also controlled [27,35,7,25,5,37]. In this paper, we take a step further to this goal under poisoning-only threat model.…”
Section: Discussionmentioning
confidence: 94%
“…However, this work assumes a much stronger threat model where adversaries not only control the training data but also control the whole training process -thus they can directly encode the latent indistinguishability requirement into the training objectives of the attacked models. Several more recent work [35,7,25,5,37] that also study this problem all follow the same threat model to Shokri et al [27]. Perhaps, a more relevant work is Tang et al [31], which points out that their source-specific poisoning attack (see Figure 1e) can reduce latent separability.…”
Section: Background and Related Workmentioning
confidence: 97%
See 1 more Smart Citation
“…Answers to this question depend on specific threat models and defensive settings we consider. Under a strong threat model where adversaries can fully control the training process, a series of recent work [36,48,9,32,5,52] show that the latent representations of poison and clean samples can be made indistinguishable by explicitly encoding the indistinguishability objective into the training loss of the backdoored model.…”
Section: B2 Adaptive Backdoor Poisoning Attacksmentioning
confidence: 99%
“…It is worth noting that GRASP represents a different type of backdoor attack compared with the stealthy backdoors proposed recently (e.g., [12] [13]). Existing stealthy backdoor methods attempt to devise specific triggers, often dependent on the target neural network model so that they are hard to detect and mitigate by defense methods.…”
Section: Introductionmentioning
confidence: 99%