2018 IEEE Security and Privacy Workshops (SPW) 2018
DOI: 10.1109/spw.2018.00014
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial Examples for Generative Models

Abstract: We explore methods of producing adversarial examples on deep generative models such as the variational autoencoder (VAE) and the VAE-GAN. Deep learning architectures are known to be vulnerable to adversarial examples, but previous work has focused on the application of adversarial examples to classification tasks. Deep generative models have recently become popular due to their ability to model input data distributions and generate realistic examples from those distributions. We present three classes of attack… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

3
146
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 190 publications
(149 citation statements)
references
References 9 publications
3
146
0
Order By: Relevance
“…However, it is reported in [128] that autoencoders seem to be much more robust to adversarial attacks than the typical classifier networks. Kos et al [121] also explored methods for computing adversarial examples for deep generative models, e.g. variational autoencoder (VAE) and the VAE-Generative Adversarial Networks (VAE-GANs).…”
Section: Attacks On Autoencoders and Generative Modelsmentioning
confidence: 99%
“…However, it is reported in [128] that autoencoders seem to be much more robust to adversarial attacks than the typical classifier networks. Kos et al [121] also explored methods for computing adversarial examples for deep generative models, e.g. variational autoencoder (VAE) and the VAE-Generative Adversarial Networks (VAE-GANs).…”
Section: Attacks On Autoencoders and Generative Modelsmentioning
confidence: 99%
“…Existing work on adversarial examples has focused largely on the space of images, be it image classification [40], generative models on images [26], image segmentation [1], face detection [37], or reinforcement learning by manipulating the images the RL agent sees [6,21]. In the discrete domain, there has been some study of adversarial examples over text classification [23] and malware classification [16,20].…”
Section: Introductionmentioning
confidence: 99%
“…A classic example is an adversary attaching a small, human-imperceptible sticker onto a stop sign that causes a self-driving car to recognize it as a yield sign. Adversarial examples have also been demonstrated in domains such as reinforcement learning [32] and generative models [31].…”
Section: Introductionmentioning
confidence: 99%