2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01616
|View full text |Cite
|
Sign up to set email alerts
|

Rethinking the Backdoor Attacks’ Triggers: A Frequency Perspective

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
60
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 107 publications
(61 citation statements)
references
References 12 publications
1
60
0
Order By: Relevance
“…In this paper, we focus on the poisoning-based backdoor attack towards image classification, where attackers can only modify the dataset instead of other training components (e.g., training loss). This threat could also happen in other tasks (Xiang et al, 2021;Zhai et al, 2021;Li et al, 2022) and with different attacker's capacities (Nguyen & Tran, 2020;Tang et al, 2020;Zeng et al, 2021a), which are out-of-scope of this paper. In general, existing attacks can be divided into two main categories based on the property of target labels, as follows:…”
Section: Backdoor Attackmentioning
confidence: 96%
See 1 more Smart Citation
“…In this paper, we focus on the poisoning-based backdoor attack towards image classification, where attackers can only modify the dataset instead of other training components (e.g., training loss). This threat could also happen in other tasks (Xiang et al, 2021;Zhai et al, 2021;Li et al, 2022) and with different attacker's capacities (Nguyen & Tran, 2020;Tang et al, 2020;Zeng et al, 2021a), which are out-of-scope of this paper. In general, existing attacks can be divided into two main categories based on the property of target labels, as follows:…”
Section: Backdoor Attackmentioning
confidence: 96%
“…Currently, there are also some approaches to alleviate the backdoor threat. Existing defenses are mostly empirical, which can be divided into five main categories, including (1) detection-based defenses (Xu et al, 2021;Zeng et al, 2021a;Xiang et al, 2022), (2) preprocessing based defenses (Doan et al, 2020;Li et al, 2021b;Zeng et al, 2021b), (3) model reconstruction based defenses (Zhao et al, 2020a;Li et al, 2021a;Zeng et al, 2022), (4) trigger synthesis based defenses Dong et al, 2021;Shen et al, 2021), and (5) poison suppression based defenses Borgnia et al, 2021). Specifically, detection-based defenses examine whether a suspicious DNN or sample is attacked and it will deny the use of malicious objects; Preprocessing based methods intend to damage trigger patterns contained in attack samples to prevent backdoor activation by introducing a preprocessing module before feeding images into DNNs; Model reconstruction based ones aim at removing the hidden backdoor in DNNs by modifying models directly; The fourth type of defenses synthesize potential trigger patterns at first, following by the second stage that the hidden backdoor is eliminated by suppressing their effects; The last type of methods depress the effectiveness of poisoned samples during the training process to prevent the creation of hidden backdoors.…”
Section: Backdoor Defensementioning
confidence: 99%
“…The detection can be performed at the granularity of individual examples (i.e., whether a given input contains a trigger), datasets (i.e., whether a subpopulation has been poisoned), or trained models (i.e., whether a given classifier contains a backdoor). To detect triggered inputs, researchers have proposed methods based on anomalous activation patterns in deep neural network layers [64], using feature attribution schemes [21,33], analyzing the prediction entropy of mixed input samples [27], or looking for high-frequency artifacts in inputs [76]. For training data inspection, Activation Clustering (AC) [14] and Spectral Signatures [65] can be used to detect different patterns of clean and poisoning samples.…”
Section: Backdoor Defensesmentioning
confidence: 99%
“…A backdoored model will behave normally on benign samples whereas constantly predicts the target label whenever the trigger appears. Currently, most existing backdoor attacks (Gu et al, 2019;Zeng et al, 2021a;Li et al, 2021c) are designed for image classification tasks and targeted towards an adversary-specified label. Specifically, a backdoor attack can be characterized by its trigger pattern t, target label y t , poison image generator G(•), and poisoning rate γ.…”
Section: Backdoor Attackmentioning
confidence: 99%