2019
DOI: 10.18637/jss.v091.i03
|View full text |Cite
|
Sign up to set email alerts
|

sgmcmc: An R Package for Stochastic Gradient Markov Chain Monte Carlo

Abstract: This paper introduces the R package sgmcmc; which can be used for Bayesian inference on problems with large datasets using stochastic gradient Markov chain Monte Carlo (SGMCMC). Traditional Markov chain Monte Carlo (MCMC) methods, such as Metropolis-Hastings, are known to run prohibitively slowly as the dataset size increases. SGMCMC solves this issue by only using a subset of data at each iteration. SGMCMC requires calculating gradients of the log likelihood and log priors, which can be time consuming and err… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
16
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
1

Relationship

3
2

Authors

Journals

citations
Cited by 10 publications
(16 citation statements)
references
References 6 publications
0
16
0
Order By: Relevance
“…This simple argument suggests that, for the same level of accuracy, we can reduce the computational cost of SGLD by O(N) if we use control variates. This is supported by a number of theoretical results (e.g., Nagapetyan et al 2017;Brosse, Durmus, and Moulines 2018;Baker et al 2019a) which show that, if we ignore the preprocessing cost of findingθ , the computational cost per-effective sample size of SGLD with control variates has a computational cost that is O(1), rather than the O(N) for SGLD with the simple gradient estimator (4).…”
Section: Estimating the Gradientmentioning
confidence: 64%
See 3 more Smart Citations
“…This simple argument suggests that, for the same level of accuracy, we can reduce the computational cost of SGLD by O(N) if we use control variates. This is supported by a number of theoretical results (e.g., Nagapetyan et al 2017;Brosse, Durmus, and Moulines 2018;Baker et al 2019a) which show that, if we ignore the preprocessing cost of findingθ , the computational cost per-effective sample size of SGLD with control variates has a computational cost that is O(1), rather than the O(N) for SGLD with the simple gradient estimator (4).…”
Section: Estimating the Gradientmentioning
confidence: 64%
“…The intuition behind this idea is that if each u i (θ) ≈ ∇U i (θ ), then this estimator can have a much smaller variance. Recent works-for example, Baker et al (2019a) and Huggins and Zou (2017) (see Bardenet, Doucet, and Holmes 2017;Bierkens, Fearnhead, and Roberts 2019;Pollock et al 2020, for similar ideas used in different Monte Carlo procedures)-have implemented this control variate technique with each u i (θ ) set as a constant. These approaches propose (i) using SGD to find an approximation to the mode of the distribution we are sampling from, which we denote asθ ; and (ii) set u i (θ ) = ∇U i (θ ).…”
Section: Estimating the Gradientmentioning
confidence: 99%
See 2 more Smart Citations
“…There is a variety of software available implementing these methods (Tran et al, 2016;Baker et al, 2016). In particular Baker et al (2016) implements the control variate methodology we discuss in this article. This paper investigates stochastic gradient Langevin dynamics (SGLD), a popular SGMCMC algorithm that discretises the Langevin diffusion.…”
Section: Introductionmentioning
confidence: 99%