2019
DOI: 10.1016/j.knosys.2018.11.002
|View full text |Cite
|
Sign up to set email alerts
|

Understanding and comparing scalable Gaussian process regression for big data

Abstract: As a non-parametric Bayesian model which produces informative predictive distribution, Gaussian process (GP) has been widely used in various fields, like regression, classification and optimization. The cubic complexity of standard GP however leads to poor scalability, which poses challenges in the era of big data. Hence, various scalable GPs have been developed in the literature in order to improve the scalability while retaining desirable prediction accuracy. This paper devotes to investigating the methodolo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
9
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
8
1

Relationship

3
6

Authors

Journals

citations
Cited by 24 publications
(9 citation statements)
references
References 29 publications
(58 reference statements)
0
9
0
Order By: Relevance
“…For example, though showing a remarkable complexity of O(m 3 ), we cannot expect the subset-of-data to perform well with increasing n. In terms of model capability, global approximations are capable of capturing the global patterns (long-term spatial correlations) but often filter out the local patterns due to the limited global inducing set. In contrast, due to the local nature, local approximations favor capturing local patterns (non-stationary features), enabling them to outperform global approximations for complicated tasks, see the solar example in [38]. The drawback however is that they ignore the global patterns to risk discontinuous predictions and local over-fitting.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, though showing a remarkable complexity of O(m 3 ), we cannot expect the subset-of-data to perform well with increasing n. In terms of model capability, global approximations are capable of capturing the global patterns (long-term spatial correlations) but often filter out the local patterns due to the limited global inducing set. In contrast, due to the local nature, local approximations favor capturing local patterns (non-stationary features), enabling them to outperform global approximations for complicated tasks, see the solar example in [38]. The drawback however is that they ignore the global patterns to risk discontinuous predictions and local over-fitting.…”
Section: Introductionmentioning
confidence: 99%
“…When sharing hyperparameters, the local structure itself may have good estimations of hyperparameter to capture some kind of local patterns[38] 22. Wang et al[157] successfully trained a MVM-based exact GP over a million data points in three days through eight GPUs.…”
mentioning
confidence: 99%
“…When GPR is applied to practical problems, GPR can give a confidence interval while outputting the mean value, making the validity of the prediction result continuously enhanced. In addition, because the GPR can quantitatively model Gaussian noise, it has excellent prediction accuracy [48,49]. Because of its good predictive ability, GPR has been widely used in data-driven modeling of various problems in industry [50][51][52][53], so GPR has also become an optional scheme in this paper.…”
Section: Gaussian Process Regression (Gpr)mentioning
confidence: 99%
“…Recently, there has been an increasing trend on the development of scalable GPs, which are classified into two core categories: global approximation and local approximation [25]. As the representative of global approximation, sparse approximation considers m (m ≪ n) global inducing pairs {X m , f m } to optimally summarize the training data by approximating the prior [26] or the posterior [27], resulting in the complexity of O(nm 2 ).…”
Section: Introductionmentioning
confidence: 99%