2018
DOI: 10.1007/s13171-018-0153-7
|View full text |Cite
|
Sign up to set email alerts
|

Subsampling MCMC - an Introduction for the Survey Statistician

Abstract: The rapid development of computing power and efficient Markov Chain Monte Carlo (MCMC) simulation algorithms have revolutionized Bayesian statistics, making it a highly practical inference method in applied work. However, MCMC algorithms tend to be computationally demanding, and are particularly slow for large datasets. Data subsampling has recently been suggested as a way to make MCMC methods scalable on massively large data, utilizing efficient sampling schemes and estimators from the survey sampling literat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 15 publications
(10 citation statements)
references
References 47 publications
0
10
0
Order By: Relevance
“…It would, thus, be interesting to extend the subsampling ideas put forth in this article to that context by combining them with INLA. Another interesting idea for future research is to apply the proposed methods of Quiroz et al (2018) when doing the final calculation of the deviance. This would further improve the speed of the only step in the current algorithm that depends on the full set of data.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…It would, thus, be interesting to extend the subsampling ideas put forth in this article to that context by combining them with INLA. Another interesting idea for future research is to apply the proposed methods of Quiroz et al (2018) when doing the final calculation of the deviance. This would further improve the speed of the only step in the current algorithm that depends on the full set of data.…”
Section: Discussionmentioning
confidence: 99%
“…This would further improve the speed of the only step in the current algorithm that depends on the full set of data. Yet, the methods from Quiroz et al (2018) would break the regularity conditions of our Theorems 1 and 2, which means that additional theoretical studies would also be required to be able to incorporate it into our framework.…”
Section: Discussionmentioning
confidence: 99%
“…As a result, the samples generated with this algorithm have a smaller size than the initial dataset X n . As the bootstrap algorithm, this one has been widely studied and used in the most different fields, from genomics [51,52] to survey science [53,54], finance [55,56] and, of course, statistics [26,57,58]. The two reviews [26,59] can be consulted by the interested reader.…”
Section: Subsamplingmentioning
confidence: 99%
“…However, learning from training data to improve inference on future data with a hierarchical model is computationally inefficient -if a new box is presented, one has to add observations of the new box to the previously available data and re-run inference on the extended data set. Inference performance can be improved by employing data subsampling [Bardenet et al 2014[Bardenet et al , 2017Korattikara et al 2014;Maclaurin and Adams 2014;Quiroz et al 2018], but the whole training data set still needs to be kept and made accessible to the inference algorithm. A hierarchical model cannot 'compress', or summarize, training data for efficient inference on future observations.…”
Section: Introductionmentioning
confidence: 99%