2019 International Conference on Robotics and Automation (ICRA) 2019
DOI: 10.1109/icra.2019.8793750
|View full text |Cite
|
Sign up to set email alerts
|

Simulating Emergent Properties of Human Driving Behavior Using Multi-Agent Reward Augmented Imitation Learning

Abstract: Recent developments in multi-agent imitation learning have shown promising results for modeling the behavior of human drivers. However, it is challenging to capture emergent traffic behaviors that are observed in real-world datasets. Such behaviors arise due to the many local interactions between agents that are not commonly accounted for in imitation learning. This paper proposes Reward Augmented Imitation Learning (RAIL), which integrates reward augmentation into the multi-agent imitation learning framework … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
30
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 41 publications
(31 citation statements)
references
References 14 publications
0
30
0
Order By: Relevance
“…This, however, is required for scenariobased verification methods. On the other hand, inverse reinforcement learning and imitation learning methods, such as GAIL [4] and variations thereof [5], [6], have been designed to actually control an automated vehicle and, thus, are capable of both, to comply with a scenario script book and to realistically mimic other road users. However, so far, these methods have mostly been tested on the NGSIM dataset [17], which comes with a limited variation of scenarios.…”
Section: B Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…This, however, is required for scenariobased verification methods. On the other hand, inverse reinforcement learning and imitation learning methods, such as GAIL [4] and variations thereof [5], [6], have been designed to actually control an automated vehicle and, thus, are capable of both, to comply with a scenario script book and to realistically mimic other road users. However, so far, these methods have mostly been tested on the NGSIM dataset [17], which comes with a limited variation of scenarios.…”
Section: B Related Workmentioning
confidence: 99%
“…However, so far, these methods have mostly been tested on the NGSIM dataset [17], which comes with a limited variation of scenarios. Furthermore, several shortcomings, like performance issues or difficulties to work for multiple agents in parallel have been reported [6]. The main goal of DeepSIL is to provide a realistically simulated environment to evaluate a motion planning software under test (MPUT).…”
Section: B Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Such non-reactive or handcrafted agents are unrealistic, and therefore create a big gap between simulators and realworld scenarios. On the other side, imitation learning (IL) techniques [5], [6], [7], [8], [9] try to "mimic" humans from the existing traffic datasets, by learning actions or reward functions instead of directly enforcing some basic safety properties of the road users (e.g. avoiding collisions, staying within lanes and below speed limits).…”
Section: Introductionmentioning
confidence: 99%