2020
DOI: 10.1287/opre.2019.1911
|View full text |Cite
|
Sign up to set email alerts
|

Simple Bayesian Algorithms for Best-Arm Identification

Abstract: This paper considers the optimal adaptive allocation of measurement effort for identifying the best among a finite set of options or designs. An experimenter sequentially chooses designs to measure and observes noisy signals of their quality with the goal of confidently identifying the best design after a small number of measurements. This paper proposes three simple and intuitive Bayesian algorithms for adaptively allocating measurement effort, and formalizes a sense in which these seemingly naive rules are t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
196
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 122 publications
(206 citation statements)
references
References 56 publications
4
196
0
Order By: Relevance
“…In essence, then, Thompson sampling with empirical Bayes shrinkage behaves similarly to algorithms like "top two Thompson sampling" (Russo, 2016). In this algorithm, Thompson sampling incorporates information about both of the top two performing arms in its construction of a posterior of the probability of an arm lying in the top two ranks.…”
Section: Dynamic Simulationsmentioning
confidence: 99%
“…In essence, then, Thompson sampling with empirical Bayes shrinkage behaves similarly to algorithms like "top two Thompson sampling" (Russo, 2016). In this algorithm, Thompson sampling incorporates information about both of the top two performing arms in its construction of a posterior of the probability of an arm lying in the top two ranks.…”
Section: Dynamic Simulationsmentioning
confidence: 99%
“…Many algorithms have been developed for R&S, including procedural approaches (Rinott, 1978), algorithms meant for solving large-scale R&S problems using fully sequential procedures (Luo, Hong, Nelson, & Wu, 2015), optimal sampling algorithm (Hunter & Pasupathy, 2013), optimal budget allocation methods (Chen, Donohue, Yücesan, & Lin, 2003), and so on. In this study, we use six different algorithms as benchmarks to compare against our proposed policy; namely, knowledge gradient (KG) by Frazier, Powell, and Dayanik (2008), top-two Thompson sampling (TTTS) by Russo (2016) which promotes more exploration compared to Thompson (posterior) sampling (TS), the most starving sequential optimal computing budget allocation (OCBA) by Chen and Lee (2011) which is modified for the Bayesian setting, general Bayesian budget allocation (GBBA) by Peng, Chen, Fu, and Hu (2016) which uses a randomized allocation function for OCBA, and approximately optimal allocation policy (AOAP) by Peng, Chong, Chen, and Fu (2018) which allocates samples in a myopic sense with the goal of minimizing the largest coefficient of variation of the difference between the best alternative (with the largest posterior mean) and the rest.…”
Section: Ranking and Selectionmentioning
confidence: 99%
“…Additionally, at the time an arm is recommended, the posteriors contain valuable information that can be used to formulate a variety of statistics helpful to assist decision makers. We consider two state-of-the-art Bayesian algorithms: BayesGap [33] and Top-two Thompson sampling [51]. For Top-two Thompson sampling, we derive a statistic based on the posteriors to inform the decision makers about the confidence of an arm recommendation: the probability of success.…”
Section: Best-arm Identification With a Fixed Budgetmentioning
confidence: 99%