2018
DOI: 10.48550/arxiv.1811.03733
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Universal Decision-Based Black-Box Perturbations: Breaking Security-Through-Obscurity Defenses

Abstract: We study the problem of finding a universal (image-agnostic) perturbation to fool machine learning (ML) classifiers (e.g., neural nets, decision tress) in the hard-label black-box setting. Recent work in adversarial ML in the white-box setting (model parameters are known) has shown that many state-of-the-art image classifiers are vulnerable to universal adversarial perturbations: a fixed human-imperceptible perturbation that, when added to any image, causes it to be misclassified with high probability Kurakin … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
2
1

Relationship

3
0

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 12 publications
0
3
0
Order By: Relevance
“…However, a model's confidence score alone may not be very reliable. For example, in computer vision, well-crafted perturbations to images can cause classifiers to make mistakes (such as, identifying a panda as a gibbon or confusing a cat with a computer) with very high confidence 5 . As we will show later, this problem also persists in the Materials Informatics pipeline (especially with distributional skewness).…”
Section: Better Evaluation and Uncertainty Quantification Techniques ...mentioning
confidence: 99%
“…However, a model's confidence score alone may not be very reliable. For example, in computer vision, well-crafted perturbations to images can cause classifiers to make mistakes (such as, identifying a panda as a gibbon or confusing a cat with a computer) with very high confidence 5 . As we will show later, this problem also persists in the Materials Informatics pipeline (especially with distributional skewness).…”
Section: Better Evaluation and Uncertainty Quantification Techniques ...mentioning
confidence: 99%
“…Universal perturbations: MimicGAN also provides effective defense against universal perturbations [41,3], which belong to the class of image-agnostic perturbations where an attack is just a single vector which when added to the entire dataset can fool a classifier. To test this defense, we first design a targeted universal perturbation using the Fast Gradient Sign Method (FGSM) [21], by computing the mean adversarial perturbation from N = 15 test images, i.e.…”
Section: Adversarial Defensementioning
confidence: 99%
“…(a) Universal Perturbations [19] Attack No Defense Cowboy [26] Defense GAN [25] MimicGAN (Ours) DeepFool [20] Universal perturbations: MimicGAN provides a natural defense against universal perturbations [19,1], which are a class of image-agnostic perturbations where an attack is just a single vector which when added to the entire dataset can fool a classifier. To test this defense, we first design a targeted universal perturbation using the Fast Gradient Sign Method (FGSM) [11], by computing the mean adversarial perturbation from N = 15 test images, i.e.…”
Section: Adversarial Defensementioning
confidence: 99%