2019 IEEE Intelligent Transportation Systems Conference (ITSC) 2019
DOI: 10.1109/itsc.2019.8917242
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive Stress Testing with Reward Augmentation for Autonomous Vehicle Validatio

Abstract: Determining possible failure scenarios is a critical step in the evaluation of autonomous vehicle systems. Real world vehicle testing is commonly employed for autonomous vehicle validation, but the costs and time requirements are high. Consequently, simulation driven methods such as Adaptive Stress Testing (AST) have been proposed to aid in validation. AST formulates the problem of finding the most likely failure scenarios as a Markov decision process, which can be solved using reinforcement learning. In pract… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
64
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 73 publications
(64 citation statements)
references
References 12 publications
0
64
0
Order By: Relevance
“…The advantage of Reinforcement Learning is that it can even change the time signals during run-time based on the assessment results in the current time step within the scenario execution. Corso et al [128] develop the approach further with a reward-augmentation technique. Both papers build on a predecessor paper [129] from the avionic domain.…”
Section: Simulation-based Falsificationmentioning
confidence: 99%
“…The advantage of Reinforcement Learning is that it can even change the time signals during run-time based on the assessment results in the current time step within the scenario execution. Corso et al [128] develop the approach further with a reward-augmentation technique. Both papers build on a predecessor paper [129] from the avionic domain.…”
Section: Simulation-based Falsificationmentioning
confidence: 99%
“…DRL has shown state-of-the-art results in playing Atari games (Mnih et al, 2015), playing chess (Silver et al, 2017), and robot manipulation from camera input (Gu et al, 2017). In recent years, different DRL techniques have been applied to falsification and most-likely failure analysis (Akazaki et al, 2018;Behzadan & Munir, 2019;Corso et al, 2019;Koren et al, 2018;Kuutti et al, 2020;Qin et al, 2019).…”
Section: Deep Reinforcement Learningmentioning
confidence: 99%
“…Meanwhile, falsification-based evaluation methods attempt to generate initial conditions or POV behaviors that force the VUT to violate the safety requirements with limited test runs. Corner cases have been generated using simulated annealing [23], rapid-exploring random tree (RRT) [24], evolutionary algorithms [25], Bayesian optimization [26], adaptive sampling [27], reinforcement learning [28], [29], etc. In [23], [24], the falsifying POVs might behave adversarially, which is not a reasonable representation of real-driving situations.…”
Section: A Scenario-based Safety Evaluation For Havsmentioning
confidence: 99%