2020
DOI: 10.48550/arxiv.2006.09255
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Corralling Stochastic Bandit Algorithms

Abstract: We study the problem of corralling stochastic bandit algorithms, that is combining multiple bandit algorithms designed for a stochastic environment, with the goal of devising a corralling algorithm that performs almost as well as the best base algorithm. We give two general algorithms for this setting, which we show benefit from favorable regret guarantees. We show that the regret of the corralling algorithms is no worse than that of the best algorithm containing the arm with the highest reward, and depends on… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
10
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(10 citation statements)
references
References 2 publications
0
10
0
Order By: Relevance
“…Thus if A J outperforms this bound and obtains logarithmic regret, our combiner algorithm will also obtain logarithmic regret, which is not obviously possible using techniques based on the Corral algorithm [8]. Note that this result also appears to improve upon [14] (Theorem 4.2) by removing a log(T ) factor, but this is because we have assumed knowledge of the time horizon T in order to set C i .…”
Section: Gap-dependent Regret Boundsmentioning
confidence: 89%
See 3 more Smart Citations
“…Thus if A J outperforms this bound and obtains logarithmic regret, our combiner algorithm will also obtain logarithmic regret, which is not obviously possible using techniques based on the Corral algorithm [8]. Note that this result also appears to improve upon [14] (Theorem 4.2) by removing a log(T ) factor, but this is because we have assumed knowledge of the time horizon T in order to set C i .…”
Section: Gap-dependent Regret Boundsmentioning
confidence: 89%
“…where the third inequality is from equation (13), and the last inequality follows from Lemma 8 Now, t|it=i 2β a t (M i T (i,t)−1 ) −1 a t ≤ C i T (i, T ) α i , where the inequality follows from the second condition for being in I t (14).…”
Section: End If End Formentioning
confidence: 97%
See 2 more Smart Citations
“…The problem of online model selection for bandit algorithms has received a lot of recent attention, as witnessed by a flurry of recent works (e.g., Agarwal et al [2017], Foster et al [2019], Chatterji et al [2020], , Arora et al [2020], , Foster et al [2020], Lee et al [2020], Bibaut et al [2020], Ghosh et al [2020]).…”
Section: Related Work and Our Contributionmentioning
confidence: 99%