2022
DOI: 10.1007/978-3-031-13185-1_10
|View full text |Cite
|
Sign up to set email alerts
|

Trainify: A CEGAR-Driven Training and Verification Framework for Safe Deep Reinforcement Learning

Abstract: Deep Reinforcement Learning (DRL) has demonstrated its strength in developing intelligent systems. These systems shall be formally guaranteed to be trustworthy when applied to safety-critical domains, which is typically achieved by formal verification performed after training. This train-then-verify process has two limits: (i) trained systems are difficult to formally verify due to their continuous and infinite state space and inexplicable AI components (i.e., deep neural networks), and (ii) the ex post facto … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(5 citation statements)
references
References 44 publications
0
4
0
Order By: Relevance
“…Optimizing the incremental model building process of COOL-MC can increase the number of supported CMARL agents. Incorporating safe CMARL approaches would also be valuable extensions to our method, as already done in the single RL domain (Carr et al 2023;Jin et al 2022;Jothimurugan et al 2022;Jansen et al 2020).…”
Section: Discussionmentioning
confidence: 99%
“…Optimizing the incremental model building process of COOL-MC can increase the number of supported CMARL agents. Incorporating safe CMARL approaches would also be valuable extensions to our method, as already done in the single RL domain (Carr et al 2023;Jin et al 2022;Jothimurugan et al 2022;Jansen et al 2020).…”
Section: Discussionmentioning
confidence: 99%
“…We presented the tool COOL-MC, which provides a tight interaction between model checking and reinforcement learning. In the future, we will extend the tool to directly incorporate safe reinforcement learning approaches [22,[25][26][27] and will extend the model expressivity to partially observable MDPs [29].…”
Section: Discussionmentioning
confidence: 99%
“…An Analytic Approach We propose an analytical approach for the control policies that are trained on discretized abstract states. In previous work (Jin et al 2022;Tian et al 2022;Li et al 2022), a compact but infinitely continuous state space S was discretized to a finite set of abstract states, i.e., S = L i=1 S i and ∀i j, S i ∩S j = ∅. Then a neural network policy π was trained on the set of abstract states.…”
Section: Reward Martingale Validationmentioning
confidence: 99%
“…If policies are implemented by DNNs on infinite and continuous state space, we take advantage of the overapproximation-based method (Lechner et al 2022). We also propose an analytical method to compute expected values precisely when policies are trained on discretized abstract state space in recently emerging approaches (Jin et al 2022;Li et al 2022;Drews, Albarghouthi, and D'Antoni 2020).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation