2020 IEEE International Conference on Image Processing (ICIP) 2020
DOI: 10.1109/icip40778.2020.9191063
|View full text |Cite
|
Sign up to set email alerts
|

Substitute Model Generation for Black-Box Adversarial Attack Based on Knowledge Distillation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 12 publications
0
8
0
Order By: Relevance
“…With the in‐depth study of black‐box attacks on deep learning models, substitute‐based adversarial attacks have received a lot of attention. Cui et al 61 proposed an algorithm to generate the substitute model of CNN models by using knowledge distillation and boosting the attacking success rate by 20%. Gao et al 62 integrated linear augmentation into substitute training and achieved success rates of 97.7% and 92.8% in MNIST and GTSRB classifiers.…”
Section: Related Workmentioning
confidence: 99%
“…With the in‐depth study of black‐box attacks on deep learning models, substitute‐based adversarial attacks have received a lot of attention. Cui et al 61 proposed an algorithm to generate the substitute model of CNN models by using knowledge distillation and boosting the attacking success rate by 20%. Gao et al 62 integrated linear augmentation into substitute training and achieved success rates of 97.7% and 92.8% in MNIST and GTSRB classifiers.…”
Section: Related Workmentioning
confidence: 99%
“…[30] illustrates that deep neural networks are susceptible to adversarial perturbations. Subsequently, more and more works [2][3][4][5]7,8,11,17,19,21,24,34] focus on the adversarial example generation task. In general, the attack task can be divided into white-box and black-box attacks, the former one can know the knowledge of the structure and parameters of the target model, and the latter one only has the access to the simple output of the target.…”
Section: Related Workmentioning
confidence: 99%
“…Most white-box algorithms [2,5,8,17,21,24] generate adversarial examples based on the gradient of loss function with respect to the inputs. For the black-box attack, some methods [3,4,7] iteratively query the outputs of target model and estimate the gradient of target model via training a substitute model; and others [11,19,34] focus on improving the transferability of adversarial examples across different models. In this work, we focus on the more practical and challenging scenario, i.e., the data-free black-box attack, which attacks the black-box target model without the need for any real data samples.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In query-based attacks [6,7,8], only the classification results or probability distribution information can be obtained by attackers. In transfer-based attacks [9,10,11], attackers can transfer part of queries from the black-box model to the local agent model selected by the attackers, in order to alle- viate the high-frequency query to the black-box model. In meta-learning-based attacks [12], attackers use the characteristics of meta-learning and knowledge distillation to make transfered simulator model more adaptable.…”
Section: Introductionmentioning
confidence: 99%