2017 IEEE International Conference on Multimedia &Amp; Expo Workshops (ICMEW) 2017
DOI: 10.1109/icmew.2017.8026257
|View full text |Cite
|
Sign up to set email alerts
|

Locally optimal detection of adversarial inputs to image classifiers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
9
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
3

Relationship

3
3

Authors

Journals

citations
Cited by 6 publications
(9 citation statements)
references
References 6 publications
0
9
0
Order By: Relevance
“…It is worth mentioning that the adversarial example detection is similar in nature to a steganalysis problem, where the digital watermarking community has developed a rich family of methods. Summarizing this experience, one can mention that to train efficient detectors of adversarial attacks, it is needed to either know a model describing an adversarial modulation along with the statistics of original data [36] or have an access to the training data sets of original and adversarial examples. Some examples of these strategies include [37][38][39][40].…”
Section: Defense Strategiesmentioning
confidence: 99%
“…It is worth mentioning that the adversarial example detection is similar in nature to a steganalysis problem, where the digital watermarking community has developed a rich family of methods. Summarizing this experience, one can mention that to train efficient detectors of adversarial attacks, it is needed to either know a model describing an adversarial modulation along with the statistics of original data [36] or have an access to the training data sets of original and adversarial examples. Some examples of these strategies include [37][38][39][40].…”
Section: Defense Strategiesmentioning
confidence: 99%
“…Current strategies to mitigate adversarial perturbations fall into two categories: (1) adversarial training algorithms to learn robust DNN classifier models [6]- [8], (2) detection algorithms to detect the adversarially perturbed inputs [9]- [11] with the classifier unchanged. While adversarial training has been successful to improve the classifier's performance against both input-dependent and universal perturbations in images, the improved robustness often come at expense of accuracy on unperturbed inputs [8].…”
Section: Introductionmentioning
confidence: 99%
“…Compared to other detectors, LO detectors are naturally suited to detection of UAPs that have small norm. LO detectors for adversarial perturbations detection were first described in [9], [10], for the setting of detecting finitely many perturbations of the input that are known to the detector. Here, we derive a locally optimal generalized likelihood ratio test (LO-GLRT) for detecting random targeted UAPs in an input of a classifier.…”
Section: Introductionmentioning
confidence: 99%
“…Current strategies to defend a classifier against adversarial perturbations fall into two categories: (1) adapting the training algorithms to learn models which are more robust, using for example adversarial training [6][7][8], (2) detect and possibly rectify the adversarially perturbed inputs [9][10][11]. While adversarial training has shown improvement of the DNN against both input dependent and input independent perturbation in images, the improved robustness often come at expense of accuracy on unperturbed inputs [8].…”
Section: Introductionmentioning
confidence: 99%
“…LO detectors are more interpretable than other detection methods for small-norm UAPs. LO detectors were earlier described in [9,10], for the setting of detecting finitely many perturbations of the input that are known to the detector. Here, we derive a locally optimal generalized likelihood ratio test (LO-GLRT) for detecting random targeted UAPs in an input of a classifier.…”
Section: Introductionmentioning
confidence: 99%