2019
DOI: 10.1109/tmm.2019.2898777
|View full text |Cite
|
Sign up to set email alerts
|

Adversarially Approximated Autoencoder for Image Generation and Manipulation

Abstract: Regularized autoencoders learn the latent codes, a structure with the regularization under the distribution, which enables them the capability to infer the latent codes given observations and generate new samples given the codes. However, they are sometimes ambiguous as they tend to produce reconstructions that are not necessarily faithful reproduction of the inputs. The main reason is to enforce the learned latent code distribution to match a prior distribution while the true distribution remains unknown. To … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
53
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
3

Relationship

4
5

Authors

Journals

citations
Cited by 92 publications
(53 citation statements)
references
References 46 publications
0
53
0
Order By: Relevance
“…Deep learning models: Inspired by the success of AlexNet [16] in the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) 2012, convolutional neural networks (CNN) have attracted a lot of attention and been successfully applied to image classification [20][21][22], object detection [4,23,24], depth estimation [25,26], image transformation [27,28], and crowd counting [29]citesajid2020plug. VGGNets [14], and GoogleNet [17], the ILSVRC winners of 2014 and 2015, proved that deeper models could significantly increase the ability of representations.…”
Section: Related Workmentioning
confidence: 99%
“…Deep learning models: Inspired by the success of AlexNet [16] in the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) 2012, convolutional neural networks (CNN) have attracted a lot of attention and been successfully applied to image classification [20][21][22], object detection [4,23,24], depth estimation [25,26], image transformation [27,28], and crowd counting [29]citesajid2020plug. VGGNets [14], and GoogleNet [17], the ILSVRC winners of 2014 and 2015, proved that deeper models could significantly increase the ability of representations.…”
Section: Related Workmentioning
confidence: 99%
“…Latent Space Exploring with GAN and VAE. Representation learning refers to the task of learning a representation of the data that can be easily exploited [29,32]. Deep probabilistic models parameterize the learned representation by a neural network.…”
Section: Related Workmentioning
confidence: 99%
“…An encoder-decoder structure was employed in [28] to learn both the generative and discriminative representations with strong supervision. In [32], the authors show that standard deep architectures can adversarially approximate to the latent space and explicitly represent factors of variation for image generation.…”
Section: Related Workmentioning
confidence: 99%
“…There are some works on making the prior more flexible through explicit parameterization [14]. In [33], the authors show that standard deep architectures can adversarially approximate to the latent space and explicitly represent factors of variation for image generation.…”
Section: Regularized Autoencodermentioning
confidence: 99%