2006 International Workshop on Integrating AI and Data Mining 2006
DOI: 10.1109/aidm.2006.4
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic Algorithm Selection Using Reinforcement Learning

Abstract: It is often the case that many algorithms exist to solve a single problem, each possessing different performance characteristics. The usual approach in this situation is to manually select the algorithm which has the best average performance. However, this strategy has drawbacks in cases where the optimal algorithm changes during an invocation of the program, in response to changes in the program's state and the computational environment. This paper presents a prototype tool that uses reinforcement learning to… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0
2

Year Published

2007
2007
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(21 citation statements)
references
References 10 publications
0
19
0
2
Order By: Relevance
“…Along these lines, Armstrong et al [1] introduce a general scheme for adaptive algorithms that employs reinforcement learning in the same manner: an 'optimization system' collects sensor data (i.e., features of the input data), invokes a reinforcement learning algorithm, and executes its decisions (i.e., which algorithm to use). The (negated) execution time of the selected algorithm serves as reward.…”
Section: Background and Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Along these lines, Armstrong et al [1] introduce a general scheme for adaptive algorithms that employs reinforcement learning in the same manner: an 'optimization system' collects sensor data (i.e., features of the input data), invokes a reinforcement learning algorithm, and executes its decisions (i.e., which algorithm to use). The (negated) execution time of the selected algorithm serves as reward.…”
Section: Background and Related Workmentioning
confidence: 99%
“…To do so, we follow [1] and introduce a preprocessor function p : Σ * → S that aggregates a trajectory τ ∈ Σ * to a state s ∈ S. As not all states of the environment are distinguishable anymore, we now deal with a partially observable Markov decision process.…”
Section: A Generic Adaptive Simulatormentioning
confidence: 99%
See 1 more Smart Citation
“…[14,15] propose 50% for the best solver, 25% for the second best, and so on. Some selection solvers [15,2] do not need a separate training phase, and performs entirely online solver selection; a weakness of this approach is that it is only possible when a large enough budget is available, so that the training phase has a minor cost. At the moment, the case of portfolios of noisy optimization solvers has not been discussed in the literature.…”
Section: Algorithm Selectionmentioning
confidence: 99%
“…This is a typical online model selection problem, for which, many algorithms have been developed and applied in a wide variety of algorithm-problem settings. In [13,1], for instance, model selection is formulated as a Markov Decision Process, and reinforcement learning is used to find the optimal algorithm scheduling strategy. In [9,6], multi-armed bandit algorithms are employed to find the optimal assignment of available computational resources.…”
Section: Introductionmentioning
confidence: 99%