2017
DOI: 10.1007/s10994-017-5663-3
|View full text |Cite
|
Sign up to set email alerts
|

Analysis of classifiers’ robustness to adversarial perturbations

Abstract: The goal of this paper is to analyze an intriguing phenomenon recently discovered in deep networks, namely their instability to adversarial perturbations (Szegedy et al, 2014). We provide a theoretical framework for analyzing the robustness of classifiers to adversarial perturbations, and show fundamental upper bounds on the robustness of classifiers. Specifically, we establish a general upper bound on the robustness of classifiers to adversarial perturbations, and then illustrate the obtained upper bound on t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

3
193
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 249 publications
(215 citation statements)
references
References 20 publications
3
193
0
Order By: Relevance
“…Moosavi-Dezfooli et al [41] prove the existence of an image-agnostic adversarial perturbation. Fawzi et al [15] extend this to theoretically show that every classifier is vulnerable to adversarial attacks. Moosavi-Dezfooli et alfurther consider the effect of the curvature of the decision boundaries on the existence of adversarial examples in [43].…”
Section: A Related Workmentioning
confidence: 94%
“…Moosavi-Dezfooli et al [41] prove the existence of an image-agnostic adversarial perturbation. Fawzi et al [15] extend this to theoretically show that every classifier is vulnerable to adversarial attacks. Moosavi-Dezfooli et alfurther consider the effect of the curvature of the decision boundaries on the existence of adversarial examples in [43].…”
Section: A Related Workmentioning
confidence: 94%
“…The robustness of image classifiers to structured and unstructured perturbations have recently attracted a lot of attention [19,16,20,3,4,12,13,14]. Despite the impressive performance of deep neural network architectures on challenging visual classification benchmarks [6,9,21,10], these classifiers were shown to be highly vulnerable to perturbations.…”
Section: Introductionmentioning
confidence: 99%
“…However, this drop is compensated by a better accommodation to deviations of the inputs, as reported by the Mean Corruption Error (MCE) scores (see [28]). Such a trade-off between accuracy and robustness has been discussed in [29]. For this experiment, we fixed k to its maximum value, d = 200, α = 2 and we used 10-NN as a classifier when using the graph smoothness loss.…”
Section: Robustnessmentioning
confidence: 99%