2021
DOI: 10.48550/arxiv.2102.10085
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Output-Weighted Sampling for Multi-Armed Bandits with Extreme Payoffs

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 0 publications
0
5
0
Order By: Relevance
“…Finally, while we focus on Bayesian experimental design, slight adjustments to the choice of acquisition function allow for Bayesian Optimization tasks [22] or for approaching other metrics of interest (e.g. mean squared error).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Finally, while we focus on Bayesian experimental design, slight adjustments to the choice of acquisition function allow for Bayesian Optimization tasks [22] or for approaching other metrics of interest (e.g. mean squared error).…”
Section: Discussionmentioning
confidence: 99%
“…where L is a diagonal matrix containing the lengthscales for each dimension and the GP hyperparameters appearing in the covariance function (σ 2 f and L in (22) are trained by maximum likelihood estimation).…”
Section: Gaussian Process Regressionmentioning
confidence: 99%
“…Although neural network architectures are attractive for approximating nonlinear regression tasks, their complexity rids them of analytical expressions. This does not allow for a traditional Bayesian treatment of uncertainty in the underlying surrogate model -a key property present for GP regression (see equation 22). Knowledge of the uncertainty of a surrogate model allows one to target model deficiencies as seen in the parameter space.…”
Section: Ensemble Of Neural Network For Uncertainty Quantificationmentioning
confidence: 99%
“…the likelihood ratio assigns to each point in the input space a measure of relevance by weighting how likely that point is to be observed "in the wild" (through the input density p b ) against its expected impact on the magnitude of the output (through the output density p µ ). As such, the likelihood ratio serves as an attention mechanism which steers the algorithm toward the extremes [31,41].…”
Section: Acquisition Functionsmentioning
confidence: 99%
“…specifically target extreme minima of the latent function. As in previous work [31,41,42], we approximate the likelihood ratio with a Gaussian mixture model in order to make the integral in (13) analytic.…”
Section: Acquisition Functionsmentioning
confidence: 99%