2019
DOI: 10.1080/02664763.2019.1572723
|View full text |Cite
|
Sign up to set email alerts
|

Parallel Markov chain Monte Carlo for Bayesian hierarchical models with big data, in two stages

Abstract: Due to the escalating growth of big data sets in recent years, new Bayesian Markov chain Monte Carlo (MCMC) parallel computing methods have been developed. These methods partition large data sets by observations into subsets. However, for Bayesian nested hierarchical models, typically only a few parameters are common for the full data set, with most parameters being groupspecific. Thus, parallel Bayesian MCMC methods that take into account the structure of the model and split the full data set by groups rather… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 29 publications
0
2
0
Order By: Relevance
“…p(data|θ) is the probability of observing "data" given θ, and is called the likelihood. The likelihood function is a function of the evidence, data, while the posterior probability is a function of the hypothesis, θ and can be expressed by the process as shown in Figure 3 [40].…”
Section: Bayesian Methodsmentioning
confidence: 99%
“…p(data|θ) is the probability of observing "data" given θ, and is called the likelihood. The likelihood function is a function of the evidence, data, while the posterior probability is a function of the hypothesis, θ and can be expressed by the process as shown in Figure 3 [40].…”
Section: Bayesian Methodsmentioning
confidence: 99%
“…Generalized Bayesian learning aims at minimizing a free energy function over a posterior distribution of the model parameters, including standard Bayesian learning as a special case (Knoblauch et al, 2019). Existing (generalized) Bayesian federated learning protocols are either based on Variational Inference (VI) (Angelino et al, 2016;Neiswanger et al, 2015;Broderick et al, 2013;Corinzia & Buhmann, 2019b) or Monte Carlo (MC) sampling Mesquita et al, 2020;Wei & Conlon, 2019). State-of-the-art methods in either category include Partitioned Variational Inference (PVI), which has been recently introduced as a unifying distributed VI framework that relies on the optimization over parametric posteriors; and Distributed Stochastic Gradient Langevin Dynamics (DSGLD), which is an MC sampling technique that maintains a number of Markov chains updated via local Stochastic Gradient Descent (SGD) with the addition of Gaussian noise Welling & Teh, 2011).…”
Section: Introductionmentioning
confidence: 99%