2016
DOI: 10.1098/rsos.160125
|View full text |Cite
|
Sign up to set email alerts
|

Fast methods for training Gaussian processes on large datasets

Abstract: Gaussian process regression (GPR) is a non-parametric Bayesian technique for interpolating or fitting data. The main barrier to further uptake of this powerful tool rests in the computational costs associated with the matrices which arise when dealing with large datasets. Here, we derive some simple results which we have found useful for speeding up the learning stage in the GPR algorithm, and especially for performing Bayesian model comparison between different covariance functions. We apply our techniques to… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
20
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 39 publications
(20 citation statements)
references
References 26 publications
0
20
0
Order By: Relevance
“…If its gradient is known and a gradient-based algorithm, such as a conjugate gradient method, can be used (as proposed in [45]), the maximisation process may be accelerated. More information about using the gradient and Hessian of Equation (21) to speed the learning phase in the GPR algorithm can be found in [28,46,47].…”
Section: Training a Gpr Modelmentioning
confidence: 99%
“…If its gradient is known and a gradient-based algorithm, such as a conjugate gradient method, can be used (as proposed in [45]), the maximisation process may be accelerated. More information about using the gradient and Hessian of Equation (21) to speed the learning phase in the GPR algorithm can be found in [28,46,47].…”
Section: Training a Gpr Modelmentioning
confidence: 99%
“…This is a major advantage for its use despite the limitations of training. Recent techniques to address the issue of training Gaussian process models for large datasets could be a way ahead in future studies [77] . Another option is to use Bayesian neural networks, rather than conventional neural networks for the choice of the surrogate model.…”
Section: Discussionmentioning
confidence: 99%
“…The covariance function, and any free parameters therein, are free to be specified; however, they can also be learnt from the training set by maximizing the probability of the training set being realized by the GP (maximizing the GP evidence). This learning process can be computationally expensive, especially for large training sets or when comparing covariance functions with many free parameters; the techniques described in [100] were used to accelerate this learning phase. The covariance functions considered here were the squared-exponential and Wendland polynomial functions used previously for waveform modeling in [43]; these covariance functions are all stationary, i.e.…”
Section: B Merger and Ringdownmentioning
confidence: 99%