2021
DOI: 10.48550/arxiv.2101.01032
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Local Black-box Adversarial Attacks: A Query Efficient Approach

Abstract: Adversarial attacks have threatened the application of deep neural networks in security-sensitive scenarios. Most existing blackbox attacks fool the target model by interacting with it many times and producing global perturbations. However, global perturbations change the smooth and insignificant background, which not only makes the perturbation more easily be perceived but also increases the query overhead. In this paper, we propose a novel framework to perturb the discriminative areas of clean examples only … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 33 publications
0
7
0
Order By: Relevance
“…Qian et al [38] proposed the CFR attack using the interpretability of neural networks and an optimization-based attack. Xiang et al [44] utilized model interpretability and a gradient-based attack to generate an initial adversarial example. Then, they generated the final example through gradient estimation and random search.…”
Section: Local Adversarial Attacksmentioning
confidence: 99%
“…Qian et al [38] proposed the CFR attack using the interpretability of neural networks and an optimization-based attack. Xiang et al [44] utilized model interpretability and a gradient-based attack to generate an initial adversarial example. Then, they generated the final example through gradient estimation and random search.…”
Section: Local Adversarial Attacksmentioning
confidence: 99%
“…As aforementioned, there have been white-box attacks or transfer-based attacks that restrict the perturbations to a small salient region. Specifically, white-box attack JSMA [26] constructs a BP-saliency map by calculating derivatives of the model output w.r.t input pixels [32], while the two transfer-based attacks [10,40] utilize CAM and Grad-CAM to extract salient regions, respectively. CAM [46] replaces the final fully connected layers with convolutional layers and global average pooling of a CNN, and localizes class-specific salient regions through forward propagation.…”
Section: Extracting Salient Regionmentioning
confidence: 99%
“…Thus, we propose the Saliency Attack, a novel black-box attack that recursively refines the perturbations in the salient region. It is worth mentioning that except white-box attack JSMA, the idea of restricting perturbations to a small region has also been implemented in transfer-based attacks [10,40], where class activation mapping (CAM) [46] and Grad-CAM [30] are adopted to generate the saliency maps. However, transfer-based attacks assume the data distribution for training the target model is available and thus could build a substitute model to approximate it, which actually belong to the grey-box setting where partial knowledge of the target model is known.…”
Section: Introductionmentioning
confidence: 99%
“…Deep models are vulnerable to adversarial examples that are maliciously constructed to mislead the models to output wrong predictions but visually indistinguishable from normal samples [182]- [185]. Adversarial training [186]- [188] is one of the most effective approaches to defend deep models against adversarial examples and enhance their robustness.…”
Section: B Collaborative Adversarial Trainingmentioning
confidence: 99%