2020 IEEE Intelligent Vehicles Symposium (IV) 2020
DOI: 10.1109/iv47402.2020.9304606
|View full text |Cite
|
Sign up to set email alerts
|

Risk-Aware High-level Decisions for Automated Driving at Occluded Intersections with Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
29
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 51 publications
(29 citation statements)
references
References 11 publications
0
29
0
Order By: Relevance
“…As proposed in [8], we can represent the whole situation at an occluded intersection using this matrix:…”
Section: Observation and State Modelmentioning
confidence: 99%
See 3 more Smart Citations
“…As proposed in [8], we can represent the whole situation at an occluded intersection using this matrix:…”
Section: Observation and State Modelmentioning
confidence: 99%
“…Reinforcement learning (RL) has gained more attention recently in order to solve complex decision making problems in robotics. Specifically for self-driving vehicles, long term optimal policies are learned using this approach for multiple scenarios such as lane changing or merging in highways [1]- [3] or yielding at unsignalized intersections [4]- [8]. Benefitting from the high representational power provided by neural networks to learn the future return of each action, learning-based policies can provide optimal behavior which is more generic and scalable compared to POMDP-based approaches like [9] and also more intelligent than rule-based approaches [10], [11].…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…They predict the future FoV of the autonomous car over the whole planning horizon and achieve a human-like driving behavior when facing occlusions. [12] proposes to learn a driving policy at unsignalized occluded intersections via reinforcement learning, which considers the future occlusions implicitly by maximizing future reward. Other approaches like [13] and [14] incorporate active exploration to eliminate future occlusion as much as possible, to actively reduce the possibility of potential hidden TPs.…”
Section: Related Workmentioning
confidence: 99%