2021
DOI: 10.48550/arxiv.2110.14468
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

DESTA: A Framework for Safe Reinforcement Learning with Markov Games of Intervention

Abstract: Exploring in an unknown system can place an agent in dangerous situations, exposing to potentially catastrophic hazards. Many current approaches for tackling safe learning in reinforcement learning (RL) lead to a trade-off between safe exploration and fulfilling the task. Though these methods possibly incur fewer safety violations, they often also lead to reduced task performance. In this paper, we take the first step in introducing a generation of RL solvers that learn to minimise safety violations while maxi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 24 publications
0
1
0
Order By: Relevance
“…A recent work [26] formulated safe RL as inference resulting in a sample efficient off-policy approach. Some approaches took a slightly different approach to formulate safety, e.g., [27] proposed a two-player framework with the cooperating task agent and safety agents, [17] proposed a safety layer that would be applied after the action is computed using a classical policy.…”
Section: Introductionmentioning
confidence: 99%
“…A recent work [26] formulated safe RL as inference resulting in a sample efficient off-policy approach. Some approaches took a slightly different approach to formulate safety, e.g., [27] proposed a two-player framework with the cooperating task agent and safety agents, [17] proposed a safety layer that would be applied after the action is computed using a classical policy.…”
Section: Introductionmentioning
confidence: 99%