2021
DOI: 10.1109/tit.2021.3081508
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Armed Bandits With Correlated Arms

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 24 publications
(17 citation statements)
references
References 21 publications
0
17
0
Order By: Relevance
“…The vaccination strategy is kept fixed over all simulation days. We define a vaccination strategy as a quintuple of the different vaccine types, divided over 5 age groups: Children (0-4), Youngsters (5-18), Young Adults (19)(20)(21)(22)(23)(24)(25), Adults (26-64) and Elderly (65+).…”
Section: Vaccine Allocationmentioning
confidence: 99%
“…The vaccination strategy is kept fixed over all simulation days. We define a vaccination strategy as a quintuple of the different vaccine types, divided over 5 age groups: Children (0-4), Youngsters (5-18), Young Adults (19)(20)(21)(22)(23)(24)(25), Adults (26-64) and Elderly (65+).…”
Section: Vaccine Allocationmentioning
confidence: 99%
“…Some algorithms are proved instance-optimal for specific interactive decision making problems. Variants of UCB algorithms are instance-optimal for bandits with various assumptions [Lattimore and Szepesvári, 2020, Gupta et al, 2021, Tirinzoni et al, 2020, Degenne et al, 2020, Magureanu et al, 2014, but are suboptimal for linear bandits [Lattimore and Szepesvari, 2017]. These algorithms rely on the optimism-in-face-of-uncertainty principle to deal with exploration-exploitation tradeoff, whereas our algorithm explicitly finds the best tradeoff.…”
Section: Additional Related Workmentioning
confidence: 99%
“…Alternatively, in a PUT that has a limit on the length of inputs to be accepted, mutations such as inserting a constant string or copying a partial byte sequence from another seed are not promising arms. While the distributions of rewards are assumed to be independent in standard stochastic bandit problems, there are studies on such problem settings where the arms are correlated, aimed at reducing regret further compared to the standard settings [30,33,55]. Even assuming independence, bandit algorithms can greatly improve the efficiency of the fuzzer.…”
Section: Correlation Of Armsmentioning
confidence: 99%