2010
DOI: 10.1007/978-3-642-13800-3_7
|View full text |Cite
|
Sign up to set email alerts
|

Algorithm Selection as a Bandit Problem with Unbounded Losses

Abstract: Algorithm selection is typically based on models of algorithm performance, learned during a separate offline training sequence, which can be prohibitively expensive. In recent work, we adopted an online approach, in which a performance model is iteratively updated and used to guide selection on a sequence of problem instances. The resulting exploration-exploitation trade-off was represented as a bandit problem with expert advice, using an existing solver for this game, but this required the setting of an arbit… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0
1

Year Published

2010
2010
2020
2020

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 26 publications
(21 citation statements)
references
References 30 publications
0
19
0
1
Order By: Relevance
“…The use of MAB algorithms to solve the EvE dilemma has been investigated in the context of selecting between different algorithm portfolios to solve decision problems [18], and in the framework of Adaptive Operator Selection by the authors. For the latter case, the Upper Confidence Bound (UCB) technique [1] was used, being referred to as the original (or basic) MAB algorithm in the following.…”
Section: Introductionmentioning
confidence: 99%
“…The use of MAB algorithms to solve the EvE dilemma has been investigated in the context of selecting between different algorithm portfolios to solve decision problems [18], and in the framework of Adaptive Operator Selection by the authors. For the latter case, the Upper Confidence Bound (UCB) technique [1] was used, being referred to as the original (or basic) MAB algorithm in the following.…”
Section: Introductionmentioning
confidence: 99%
“…For the same set-up, an allocation strategy is proposed in [16] based on updating dynamically the belief over the run-time distribution. Finally, when a set of time allocation strategies are available and the optimization problem is to be solved several times, one can use the standard multi-armed bandit framework as in [9,10,11].…”
Section: Introductionmentioning
confidence: 99%
“…Further references on algorithm selection can be found in [8,9]. Literature on parallel computing, grid computing, distributed computing [6,1,18] is focused on allocation of dynamically changing computational resources, in a transparent and fault tolerant manner.…”
Section: Related Workmentioning
confidence: 99%
“…In [9], basing on [3], we introduced EXP3LIGHT-A, a BPS which guarantees a bound on regret when the maximum loss is unknown a priori. Note that any bound on the regret of the chosen BPS will determine a bound on the regret of GAMBLETA with respect to the best time allocator.…”
Section: Algorithm 1 Gambleta(at Bps) Gambling Time Allocatormentioning
confidence: 99%
See 1 more Smart Citation