2018
DOI: 10.1007/s11222-018-9826-2
|View full text |Cite
|
Sign up to set email alerts
|

Control variates for stochastic gradient MCMC

Abstract: It is well known that Markov chain Monte Carlo (MCMC) methods scale poorly with dataset size. A popular class of methods for solving this issue is stochastic gradient MCMC (SGMCMC). These methods use a noisy estimate of the gradient of the log-posterior, which reduces the per iteration computational cost of the algorithm. Despite this, there are a number of results suggesting that stochastic gradient Langevin dynamics (SGLD), probably the most popular of these methods, still has computational cost proportional… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
95
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
8
1

Relationship

3
6

Authors

Journals

citations
Cited by 55 publications
(95 citation statements)
references
References 18 publications
0
95
0
Order By: Relevance
“…Increasing the minibatch size will reduce the variance of the gradient estimate, but increase the per iteration computational cost of the SGMCMC algorithm. Recently control variates (Ripley, 2009) have been used to reduce the variance in the gradient estimate of SGMCMC (Dubey et al, 2016;Nagapetyan et al, 2017;Baker et al, 2017). Using these improved gradient estimates have been shown to lead to improvements in the mean squared error (MSE) of the algorithm (Dubey et al, 2016), as well as its computational cost (Nagapetyan et al, 2017;Baker et al, 2017).…”
Section: Stochastic Gradient Mcmc With Control Variatesmentioning
confidence: 99%
See 1 more Smart Citation
“…Increasing the minibatch size will reduce the variance of the gradient estimate, but increase the per iteration computational cost of the SGMCMC algorithm. Recently control variates (Ripley, 2009) have been used to reduce the variance in the gradient estimate of SGMCMC (Dubey et al, 2016;Nagapetyan et al, 2017;Baker et al, 2017). Using these improved gradient estimates have been shown to lead to improvements in the mean squared error (MSE) of the algorithm (Dubey et al, 2016), as well as its computational cost (Nagapetyan et al, 2017;Baker et al, 2017).…”
Section: Stochastic Gradient Mcmc With Control Variatesmentioning
confidence: 99%
“…We implement the formulation of Baker et al (2017), who replace the gradient estimate ∇ θ log p(θ|x) with…”
Section: Stochastic Gradient Mcmc With Control Variatesmentioning
confidence: 99%
“…However, although they often work well in practice it can be difficult to know just how accurate the results are for any given application. Furthermore, many of these algorithms still have a computational cost that increases linearly with data size (Bardenet et al ., 2017; Nagapetyan et al ., 2017; Baker et al ., 2019).…”
Section: Introductionmentioning
confidence: 99%
“…Both approaches show good performance when the subset posteriors are near Gaussian, which is expected for adequately large sample sizes for each subset, based on the Bayesian central limit theorem (Bernstein von-Mises theorem; see Van der Vaart [41], and Le Cam and Yang [19]). However, for non-Gaussian posteriors, the methods may have unreliable performance (Baker et al [3]; Neiswanger et al [28]; Miroshnikov et al [25]). The method of Neiswanger et al [28] also has limitations as the number of unknown model parameters increases, since kernel density estimation becomes infeasible in larger dimensions (Wang and Dunson [42]; Scott [37]).…”
Section: Introductionmentioning
confidence: 99%