2022 American Control Conference (ACC) 2022
DOI: 10.23919/acc53348.2022.9867204
|View full text |Cite
|
Sign up to set email alerts
|

Representation Learning for Context-Dependent Decision-Making

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 15 publications
0
0
0
Order By: Relevance
“…The piece-wise stationary stochastic multi-armed bandits have also been considered for the linear bandits; the linear bandit is when the reward is the inner product of an unknown vector and an action vector chosen by the learner [8]. Prior work has considered the cases where the unknown parameter vector changes abruptly [9], and where the rewards are dependent on the previously chosen action vectors [10]. Some non-stationary methods focused on the impact of distributional change in the form of variational budget, which is a known constant that bounds the cumulative changes in the distributions [11], [12].…”
Section: Introductionmentioning
confidence: 99%
“…The piece-wise stationary stochastic multi-armed bandits have also been considered for the linear bandits; the linear bandit is when the reward is the inner product of an unknown vector and an action vector chosen by the learner [8]. Prior work has considered the cases where the unknown parameter vector changes abruptly [9], and where the rewards are dependent on the previously chosen action vectors [10]. Some non-stationary methods focused on the impact of distributional change in the form of variational budget, which is a known constant that bounds the cumulative changes in the distributions [11], [12].…”
Section: Introductionmentioning
confidence: 99%