Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.259
|View full text |Cite
|
Sign up to set email alerts
|

FlowPrior: Learning Expressive Priors for Latent Variable Sentence Models

Abstract: Variational autoencoders (VAEs) are widely used for latent variable modeling of text. We focus on variations that learn expressive prior distributions over the latent variable. We find that existing training strategies are not effective for learning rich priors, so we add the importance-sampled log marginal likelihood as a second term to the standard VAE objective to help when learning the prior. Doing so improves results for all priors evaluated, including a novel choice for sentence VAEs based on normalizing… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
9
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(10 citation statements)
references
References 21 publications
0
9
0
Order By: Relevance
“…Expressive Prior and Posterior In VAEs VAEs usually employ simple Gaussian distribution as the prior and spherical Gaussian distributions with diagonal co-variance matrices as the variational posteriors (Higgins et al, 2017;He et al, 2019;Fu et al, 2019). Such predefined forms in traditional formulations hinder VAEs from the expressivity of the model (Ding and Gimpel, 2021), thus further inducing the posterior collapse (Fang et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Expressive Prior and Posterior In VAEs VAEs usually employ simple Gaussian distribution as the prior and spherical Gaussian distributions with diagonal co-variance matrices as the variational posteriors (Higgins et al, 2017;He et al, 2019;Fu et al, 2019). Such predefined forms in traditional formulations hinder VAEs from the expressivity of the model (Ding and Gimpel, 2021), thus further inducing the posterior collapse (Fang et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…To improve the representation capacity, some efforts try more expressive priors. MoG-VAE (Ding and Gimpel, 2021) considers a uniform mixture of Gaussians as the prior, Vamp-VAE (Tomczak and Welling, 2018) introduces "Variational Mixture of Posteriors" prior (VampPrior). APo-VAE (Dai et al, 2021) adopts VampPrior to learn a hyperbolic latent space.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations