2017
DOI: 10.1007/s11222-017-9729-7
|View full text |Cite
|
Sign up to set email alerts
|

Gaussian variational approximation with sparse precision matrices

Abstract: We consider the problem of learning a Gaussian variational approximation to the posterior distribution for a high-dimensional parameter, where we impose sparsity in the precision matrix to reflect appropriate conditional independence structure in the model. Incorporating sparsity in the precision matrix allows the Gaussian variational distribution to be both flexible and parsimonious, and the sparsity is achieved through parameterization in terms of the Cholesky factor. Efficient stochastic gradient methods wh… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
42
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
1
1

Relationship

3
4

Authors

Journals

citations
Cited by 42 publications
(44 citation statements)
references
References 27 publications
0
42
0
Order By: Relevance
“…Unbiased gradient estimates can be constructed in several ways depending on whether Roeder et al (2017), Tan and Nott (2018) and Tan (2018), an advantage of estimating both terms in (8) using the same samples s ∼ φ(s) is that the stochasticity arising from s in the two terms cancel out each other resulting in smaller variation in the gradients at convergence. As log q λ (θ) depends on {µ, C} directly as well as through θ, we apply chain rule to obtain…”
Section: Stochastic Variational Inferencementioning
confidence: 99%
“…Unbiased gradient estimates can be constructed in several ways depending on whether Roeder et al (2017), Tan and Nott (2018) and Tan (2018), an advantage of estimating both terms in (8) using the same samples s ∼ φ(s) is that the stochasticity arising from s in the two terms cancel out each other resulting in smaller variation in the gradients at convergence. As log q λ (θ) depends on {µ, C} directly as well as through θ, we apply chain rule to obtain…”
Section: Stochastic Variational Inferencementioning
confidence: 99%
“…a diagonal approximation) are very different, and even a crude allowance for posterior correlation with a small number of factors can grealy improve estimation of the posterior marginal distributions. Finally, we also compare the variational marginal density of the regression coefficients with the method in Tan and Nott (2016). The method of Tan and Nott (2016) gives similar answers to MCMC in this example, as shown in Figure 5 of their manuscript, so the Tan and Nott (2016) can be considered both a gold standard for a normal approximation as well as a good gold standard more globally.…”
Section: Mixed Logistic Regressionmentioning
confidence: 83%
“…which shows that a gradient estimate based on a single sample of f using (9) and (10) will be zero at the mode. Similar points are discussed in Salimans and Knowles (2013), Han et al (2016) and Tan and Nott (2016) in other contexts and we use the gradient estimates based on (9) and (10) (6), (9) and (10) at λ (t) where the expectations are approximated from the single sample ( (t) , z (t) ).…”
Section: Stochastic Gradient Variational Bayesmentioning
confidence: 97%
See 2 more Smart Citations