2018
DOI: 10.48550/arxiv.1811.02073
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

QUOTA: The Quantile Option Architecture for Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…This requires a representation of the risks in decision making with a principled way of evaluating the risks to decide which risk level for the vehicle to undertake. Experiments by Zhang et al (2018) showed that considering no risks but only acting according to the mean of return as in traditional RL is not always the best decision and can fail to discover the optimal policy. In contrast, acting according to different levels of risks can find the optimal policy in environments with a diverse range of different transition and reward dynamics.…”
Section: Temporal Questions and Question Hierarchiesmentioning
confidence: 99%
“…This requires a representation of the risks in decision making with a principled way of evaluating the risks to decide which risk level for the vehicle to undertake. Experiments by Zhang et al (2018) showed that considering no risks but only acting according to the mean of return as in traditional RL is not always the best decision and can fail to discover the optimal policy. In contrast, acting according to different levels of risks can find the optimal policy in environments with a diverse range of different transition and reward dynamics.…”
Section: Temporal Questions and Question Hierarchiesmentioning
confidence: 99%