2016 Winter Simulation Conference (WSC) 2016
DOI: 10.1109/wsc.2016.7822140
|View full text |Cite
|
Sign up to set email alerts
|

Warm starting Bayesian optimization

Abstract: We develop a framework for warm-starting Bayesian optimization, that reduces the solution time required to solve an optimization problem that is one in a sequence of related problems. This is useful when optimizing the output of a stochastic simulator that fails to provide derivative information, for which Bayesian optimization methods are well-suited. Solving sequences of related optimization problems arises when making several business decisions using one optimization model and input data collected over diff… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
25
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
2
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 34 publications
(25 citation statements)
references
References 16 publications
0
25
0
Order By: Relevance
“…Morales-Enciso and Branke (2015) consider the optimisation of a changing objective function with EGO, and propose three different ways to re-use information gathered on previous objective functions to speed up optimisation of the current objective function: using the old posterior mean as the new prior mean, treating old samples as noisy, and treating time as an extra input dimension to the learnt function. The latter idea is also used by Poloczek et al (2016) to warm-start Bayesian Optimisation, however with the Knowledge Gradient as the acquisition function for the current optimisation task. A similar problem has been much studied in the field of Contextual multi-armed bandits, an extension of the multi-armed bandit problem where the best arm depends on a context that is randomly changing with time.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Morales-Enciso and Branke (2015) consider the optimisation of a changing objective function with EGO, and propose three different ways to re-use information gathered on previous objective functions to speed up optimisation of the current objective function: using the old posterior mean as the new prior mean, treating old samples as noisy, and treating time as an extra input dimension to the learnt function. The latter idea is also used by Poloczek et al (2016) to warm-start Bayesian Optimisation, however with the Knowledge Gradient as the acquisition function for the current optimisation task. A similar problem has been much studied in the field of Contextual multi-armed bandits, an extension of the multi-armed bandit problem where the best arm depends on a context that is randomly changing with time.…”
Section: Literature Reviewmentioning
confidence: 99%
“…In particular, it would be interesting to assess what happens when participants learn in either rougher, smoother, or dynamically changing environments, before assessing them on the same level of smoothness. This could be used to develop and test different hierarchies of function learning models, such as warm starting (i.e., adjusting hyper-parameters over environments Poloczek et al, 2016), mismatched learning (i.e., assuming a fixed length-scale that is either too small or too large for the test set), or even "learning-to-learn"-algorithms (Chen et al, 2017) within environments of the same class over time.…”
Section: Gaussian Process Learning Curves Match Perceived Predictabilitymentioning
confidence: 99%
“…In many tuning scenarios, there are correlated tuning tasks, and one can use such correlation to improve performance on each task. Multitask tuning has been used for continuous sets of tasks [39,38] and discrete tasks [31,54,44,13,35]. In the discrete case, one can use a multi-output GP model, such as the linear coregionalization model (LCM) [3,22] or the intrinsic model of coregionalization (IMC) [3,18,9], to model performance across correlated tasks [54,44,13,35].…”
Section: Introductionmentioning
confidence: 99%
“…Multitask tuning has been used for continuous sets of tasks [39,38] and discrete tasks [31,54,44,13,35]. In the discrete case, one can use a multi-output GP model, such as the linear coregionalization model (LCM) [3,22] or the intrinsic model of coregionalization (IMC) [3,18,9], to model performance across correlated tasks [54,44,13,35]. Some works further extended acquisition functions in BO to multi-task settings [31,54,44,13].…”
Section: Introductionmentioning
confidence: 99%