Proceedings of the Genetic and Evolutionary Computation Conference Companion 2021
DOI: 10.1145/3449726.3463164
|View full text |Cite
|
Sign up to set email alerts
|

How Bayesian should Bayesian optimisation be?

Abstract: Bayesian optimisation (BO) uses probabilistic surrogate modelsusually Gaussian processes (GPs) -for the optimisation of expensive black-box functions. At each BO iteration, the GP hyperparameters are fit to previously-evaluated data by maximising the marginal likelihood. However, this fails to account for uncertainty in the hyperparameters themselves, leading to overconfident model predictions. This uncertainty can be accounted for by taking the Bayesian approach of marginalising out the model hyperparameters.… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(7 citation statements)
references
References 48 publications
0
5
0
Order By: Relevance
“…Bar heights correspond to the proportion of times each model and scalarisation combination was the best on each test problem. As can be seen from the figure, MBORE with XGB and our novel PHC scalariser (15) has the best overall performance for both benchmarks. Surprisingly, MBORE with the MLP classification method performs worse than the best performing method on all the 63 WFG test problems.…”
Section: Synthetic Benchmarksmentioning
confidence: 87%
See 3 more Smart Citations
“…Bar heights correspond to the proportion of times each model and scalarisation combination was the best on each test problem. As can be seen from the figure, MBORE with XGB and our novel PHC scalariser (15) has the best overall performance for both benchmarks. Surprisingly, MBORE with the MLP classification method performs worse than the best performing method on all the 63 WFG test problems.…”
Section: Synthetic Benchmarksmentioning
confidence: 87%
“…In this work we use a Matérn 5/2 kernel, as recommended for modelling realistic functions [69,73]. The kernel's hyperparameters 𝜽 are learnt via maximising the log marginal likelihood [15,58].…”
Section: Bayesian Optimisationmentioning
confidence: 99%
See 2 more Smart Citations
“…136 Additionally, slice sampling, 137 adaptive importance sampling, 138 and entropy-based methods 139 have been used within Bayesian optimisation, where there is interest in moving beyond single-point estimates to fully-Bayesian approaches. 140,141 The main drawback with these methods is the high computational cost, and ideally they should only be employed when it will provide a substantial advantage over single point estimates. However, this condition cannot be known a priori .…”
Section: Future Workmentioning
confidence: 99%