2021
DOI: 10.1002/int.22458
|View full text |Cite
|
Sign up to set email alerts
|

Detection defense against adversarial attacks with saliency map

Abstract: It is well established that neural networks are vulnerable to adversarial examples, which are almost imperceptible on human vision and can cause the deep models misbehave. Such phenomenon may lead to severely inestimable consequences in the safety and security critical applications. Existing defenses are trend to harden the robustness of models against adversarial attacks, for example, adversarial training technology. However, these are usually intractable to implement due to the high cost of retraining and th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 20 publications
(14 citation statements)
references
References 24 publications
0
11
0
Order By: Relevance
“…Therefore, acquiring features about the classes from the differences between adversarial examples and the original images is the key to detect and correctly classify adversarial examples simultaneously. Through surveys and experiments, we observe linkage between the evolutions of the features in the CNNs and the differences of adversarial examples from original images 19,20 . Although the original images and adversarial examples are difficult to be perceived by humans, they have exhibited increasing differences in their feature maps in the feedforward process of the model.…”
Section: Introductionmentioning
confidence: 95%
See 1 more Smart Citation
“…Therefore, acquiring features about the classes from the differences between adversarial examples and the original images is the key to detect and correctly classify adversarial examples simultaneously. Through surveys and experiments, we observe linkage between the evolutions of the features in the CNNs and the differences of adversarial examples from original images 19,20 . Although the original images and adversarial examples are difficult to be perceived by humans, they have exhibited increasing differences in their feature maps in the feedforward process of the model.…”
Section: Introductionmentioning
confidence: 95%
“…Through surveys and experiments, we observe linkage between the evolutions of the features in the CNNs and the differences of adversarial examples from original images. 19,20 Although the original images and adversarial examples are difficult to be perceived by humans, they have exhibited increasing differences in their feature maps in the feedforward process of the model. We call it the Adversarial Feature Separability (AFS).…”
mentioning
confidence: 99%
“…[5][6][7][8][9][10] These perturbations are imperceptible to human beings but can easily fool DNNs, which raises invisible threats to the vision-based automatic decision. [11][12][13][14][15] Consequently, the robustness of DNNs encounters great challenges in real-world applications. 16,17 For example, the existence of AEs can pose severe security threats for traffic sign recognition in autonomous driving.…”
Section: Introductionmentioning
confidence: 99%
“…However, this also increases the risk of disinformation crimes caused by manipulations with malicious intention 14–18 . For instance, fake news with maliciously manipulated image (e.g., on COVID‐19) can cause serious negative impacts on national security and social stability 19–22 …”
Section: Introductionmentioning
confidence: 99%
“…[14][15][16][17][18] For instance, fake news with maliciously manipulated image (e.g., on COVID-19) can cause serious negative impacts on national security and social stability. [19][20][21][22] To alleviate the risks of manipulated images, various manipulation detection methods are proposed to recognize the manipulated images. [23][24][25][26] Early research solve image manipulation detection as a classification task.…”
Section: Introductionmentioning
confidence: 99%