2017 IEEE Winter Conference on Applications of Computer Vision (WACV) 2017
DOI: 10.1109/wacv.2017.131
|View full text |Cite
|
Sign up to set email alerts
|

Deep Feature Consistent Variational Autoencoder

Abstract: We present a novel method for constructing Variational Autoencoder (VAE). Instead of using pixel-by-pixel loss, we enforce deep feature consistency between the input and the output of a VAE, which ensures the VAE's output to preserve the spatial correlation characteristics of the input, thus leading the output to have a more natural visual appearance and better perceptual quality. Based on recent deep learning works such as style transfer, we employ a pre-trained deep convolutional neural network (CNN) and use… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
214
0
4

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 271 publications
(221 citation statements)
references
References 30 publications
3
214
0
4
Order By: Relevance
“…This is known as the reparameterization trick and allows the calculation of the gradient of the loss function with respect to the parameters of this architecture [56]. VAEs are known to produce blurry reconstructions [57,58] of the inputs, but more meaningful data representations, unlike vanilla AEs. That is because the VAEs are forced to minimize reconstruction errors from feature vectors sampled from the latent distribution.…”
Section: Unsupervised Learningmentioning
confidence: 99%
“…This is known as the reparameterization trick and allows the calculation of the gradient of the loss function with respect to the parameters of this architecture [56]. VAEs are known to produce blurry reconstructions [57,58] of the inputs, but more meaningful data representations, unlike vanilla AEs. That is because the VAEs are forced to minimize reconstruction errors from feature vectors sampled from the latent distribution.…”
Section: Unsupervised Learningmentioning
confidence: 99%
“…Deep Recurrent Attentive Writer (DRAW) [29] combines spatial attention mechanism with a sequential variational autoencoding framework that allows iterative generation of images. [30] and [17] consider replacing per-pixel loss with perceptual similarities using either multi-scale structural similarity score or a perceptual loss based on deep features extracted from pretrained deep networks.…”
Section: Variational Autoencodermentioning
confidence: 99%
“…Following works [19,31,32,33,34,35,36,37,38,39,40,41] [17] and Wasserstein GAN (WGAN) [19] to improve the perceptual quality of the output images generated by VAE and enhance the effectiveness of VAE representations for semi-supervised learning. In addition, a combination of VAE and GAN was also proposed by [42].…”
Section: Generative Adversarial Networkmentioning
confidence: 99%
See 2 more Smart Citations