2018
DOI: 10.48550/arxiv.1806.02867
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Direct Optimization through $\arg \max$ for Discrete Variational Auto-Encoder

Abstract: Reparameterization of variational auto-encoders with continuous random variables is an effective method for reducing the variance of their gradient estimates. Our work optimizes the discrete VAE objective directly, using its Gumbel-Max reparameterization, by applying the direct loss minimization technique to generative models. This optimization technique propagates gradients through the reparameterized arg max, which are estimated by the difference of gradients of two arg max predictions. This realization prov… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(6 citation statements)
references
References 13 publications
0
6
0
Order By: Relevance
“…This derivative requires A function evaluations, similar to RAM. Lorberbom et al (2018) extend the single variable result to multivariate distributions similar to Eqs. ( 4) and ( 5).…”
Section: Argmaxmentioning
confidence: 87%
See 3 more Smart Citations
“…This derivative requires A function evaluations, similar to RAM. Lorberbom et al (2018) extend the single variable result to multivariate distributions similar to Eqs. ( 4) and ( 5).…”
Section: Argmaxmentioning
confidence: 87%
“…Each expectation can be computed by sampling thereby trading off computational effort for increased variance. Lorberbom et al (2018) show that…”
Section: Argmaxmentioning
confidence: 99%
See 2 more Smart Citations
“…Some works rely on explicit summation of the expectation, either for the marginal distribution (Titsias & Lázaro-Gredilla, 2015) or globally summing some categories while sampling from the remainder (Liang et al, 2018;Liu et al, 2019). Other approaches use a finite difference approximation to the gradient (Lorberbom et al, 2018;. Yin et al (2019) introduced ARSM, which uses multiple model evaluations where the number adapts automatically to the uncertainty.…”
Section: Introductionmentioning
confidence: 99%