2015
DOI: 10.48550/arxiv.1508.04175
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Nonparametric Bayesian Aggregation for Massive Data

Zuofeng Shang,
Botao Hao,
Guang Cheng

Abstract: We develop a set of scalable Bayesian inference procedures for a general class of nonparametric regression models. Specifically, nonparametric Bayesian inferences are separately performed on each subset randomly split from a massive dataset, and then the obtained local results are aggregated into global counterparts. This aggregation step is explicit without involving any additional computation cost. By a careful partition, we show that our aggregated inference results obtain an oracle rule in the sense that t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2017
2017
2017
2017

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 21 publications
0
1
0
Order By: Relevance
“…We show that when k is controlled to increase in some proper order of n as n tends to infinity, the Bayes L 2 -risk of the DISK posterior achieves near minimax optimal convergence rates under different types of covariance functions. There are some theoretical results in this direction (Shang and Cheng, 2015, Cheng and Shang, 2017, Szabo and van Zanten, 2017, but DISK is the first general Bayesian framework addressing these theoretical problems with a focus on computationally efficient posterior computations in massive data applications with complex nonparametric models, while avoiding restrictive assumptions that limit wide applicability.…”
Section: Introductionmentioning
confidence: 99%
“…We show that when k is controlled to increase in some proper order of n as n tends to infinity, the Bayes L 2 -risk of the DISK posterior achieves near minimax optimal convergence rates under different types of covariance functions. There are some theoretical results in this direction (Shang and Cheng, 2015, Cheng and Shang, 2017, Szabo and van Zanten, 2017, but DISK is the first general Bayesian framework addressing these theoretical problems with a focus on computationally efficient posterior computations in massive data applications with complex nonparametric models, while avoiding restrictive assumptions that limit wide applicability.…”
Section: Introductionmentioning
confidence: 99%