2022
DOI: 10.1109/tnnls.2021.3071401
|View full text |Cite
|
Sign up to set email alerts
|

Deep Mixture Generative Autoencoders

Abstract: Variational autoencoders (VAEs) are one of the most popular unsupervised generative models which rely on learning latent representations of data. In this paper, we extend the classical concept of Gaussian mixtures into the deep variational framework by proposing a mixture of VAEs (MVAE). Each component in the MVAE model is implemented by a variational encoder and has an associated sub-decoder. The separation between the latent spaces modelled by different encoders is enforced using the d-variable Hilbert-Schmi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 28 publications
(15 citation statements)
references
References 25 publications
(48 reference statements)
0
15
0
Order By: Relevance
“…The maximum size for STM and LTM is set to 512 and 2048, respectively, to avoid increasing the computational cost. We follow the settings from [61], as described in Appendix-H.3 from SM, after resizing all images to 64 × 64 pixels. The FID and IS results are provided in Table 5 and the results of all baselines (training on a single dataset) are cited from [61].…”
Section: Evaluation Of the Reconstruction Qualitymentioning
confidence: 99%
See 1 more Smart Citation
“…The maximum size for STM and LTM is set to 512 and 2048, respectively, to avoid increasing the computational cost. We follow the settings from [61], as described in Appendix-H.3 from SM, after resizing all images to 64 × 64 pixels. The FID and IS results are provided in Table 5 and the results of all baselines (training on a single dataset) are cited from [61].…”
Section: Evaluation Of the Reconstruction Qualitymentioning
confidence: 99%
“…We follow the settings from [61], as described in Appendix-H.3 from SM, after resizing all images to 64 × 64 pixels. The FID and IS results are provided in Table 5 and the results of all baselines (training on a single dataset) are cited from [61]. The visual results are shown in Fig.…”
Section: Evaluation Of the Reconstruction Qualitymentioning
confidence: 99%
“…The primary drawback of these models is that most existing DAMs would only focus on classification tasks and can not learn meaningful representations across domains under the unsupervised learning setting. Furthermore, there were some attempts for using mixture models to learn complex datasets [38], [39], [40] or learn an infinite number of tasks [12], [18], [41], [42]. However, the expansion of these models relies on the estimation of the sample log-likelihood, which requires each expert to have an explicit probabilistic function form.…”
Section: Dynamic Architecturesmentioning
confidence: 99%
“…In this paper we consider a model made up of a mixture of networks [45] which is able to deal with three different learning scenarios: supervised, semi-supervised and unsupervised, under the lifelong learning setting. Let us consider a sequence of tasks and denote…”
Section: The Lifelong Mixture Of Vaes a Problem Formulationmentioning
confidence: 99%