2021
DOI: 10.48550/arxiv.2105.03689
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Self-Supervised Adversarial Example Detection by Disentangled Representation

Abstract: Deep learning models are known to be vulnerable to adversarial examples that are elaborately designed for malicious purposes and are imperceptible to the human perceptual system. Autoencoder, when trained solely over benign examples, has been widely used for (self-supervised) adversarial detection based on the assumption that adversarial examples yield larger reconstruction error. However, because lacking adversarial examples in its training and the too strong generalization ability of autoencoder, this assump… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 19 publications
(43 reference statements)
0
3
0
Order By: Relevance
“…Other robust ML algorithms may also be able to run over corrupted labels [49,53].The Gold Loss Correction (GLC) method in [29] utilizes a small set of trusted labels to improve the setting's accuracy. In the presence of high degrees of label corruption, the ML models overfit false samples into the corrupted labels, such as [5,28,69,70]. These algorithms could suggest extra unsupervised objective training with reliable signals.…”
Section: Self-supervised Learning (Ssl) Approachesmentioning
confidence: 99%
“…Other robust ML algorithms may also be able to run over corrupted labels [49,53].The Gold Loss Correction (GLC) method in [29] utilizes a small set of trusted labels to improve the setting's accuracy. In the presence of high degrees of label corruption, the ML models overfit false samples into the corrupted labels, such as [5,28,69,70]. These algorithms could suggest extra unsupervised objective training with reliable signals.…”
Section: Self-supervised Learning (Ssl) Approachesmentioning
confidence: 99%
“…Self-supervised learning provides a way for learning representation from unlabeled data. Recent efforts have been made toward using self-supervised algorithms in order to learn a disentangle representation [ 57 , 58 , 59 , 60 ]. However, recent studies have reported that the existing SSL methods often struggle to learn disentangled representations of the data [ 60 ].…”
Section: Introductionmentioning
confidence: 99%
“…From another line of research, adversarial examples of deep neural models have been extensively studied [25,26,23,24,27,28]. The seminal work in [25] made use of elaborately crafted adversarial perturbations, which are very tiny and unnoticeable to human eyes, to cause misclassifications of a victim model.…”
Section: Introductionmentioning
confidence: 99%