2015
DOI: 10.17706/jsw.10.3.239-249
|View full text |Cite
|
Sign up to set email alerts
|

Accelerated -Greedy Multi Armed Bandit Algorithm for Online Sequential-Selections Applications

Abstract: Current algorithms for solving multi-armed bandit (MAB) problem in stationary observations often perform well. Although this performance may be acceptable with accurate parameter settings, most of them degrade under non stationary observations. We setup an incremental ε-greedy model with stochastic mean equation as its action-value function which is more applicable to real-world problems. Unlike the iterative algorithms suffering from step size dependency, we propose an adaptive step-size model (ASM) to introd… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 13 publications
0
1
0
Order By: Relevance
“…Overview of generalized weighted averages [8]. The application of (8) in the UCB-tuned algorithm and the asymptotically optimal UCB is respectively as ( 9) and ( 10).…”
Section: Weighted Methods For Algorithm Optimizationmentioning
confidence: 99%
“…Overview of generalized weighted averages [8]. The application of (8) in the UCB-tuned algorithm and the asymptotically optimal UCB is respectively as ( 9) and ( 10).…”
Section: Weighted Methods For Algorithm Optimizationmentioning
confidence: 99%