2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2021
DOI: 10.1109/iros51168.2021.9635925
|View full text |Cite
|
Sign up to set email alerts
|

Benchmarking Safe Deep Reinforcement Learning in Aquatic Navigation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(12 citation statements)
references
References 16 publications
0
12
0
Order By: Relevance
“…We will need two learning rates, one for the actor-network (α actor ) and the other for the critic network (α critic ) because we have two types of networks. The use of the percent of times that a random action is executed is explained by equation (5).…”
Section: Ddpg + Her and Gamentioning
confidence: 99%
See 1 more Smart Citation
“…We will need two learning rates, one for the actor-network (α actor ) and the other for the critic network (α critic ) because we have two types of networks. The use of the percent of times that a random action is executed is explained by equation (5).…”
Section: Ddpg + Her and Gamentioning
confidence: 99%
“…Reinforcement Learning (RL) [1] has recently been applied to a variety of applications, including robotic table tennis [2], surgical robot planning [3], rapid motion planning in bimanual suture needle regrasping [4], and Aquatic Navigation [5]. Each one of these applications employs RL as a motivating alternative to automating manual labor.…”
Section: Introductionmentioning
confidence: 99%
“…To date, existing techniques for improving the safety of robotic systems rely mostly on Lagrangian multipliers [39], [53], [57], and do not provide formal safety guarantees, but rather optimize the training in an attempt to learn the required policies [12]. Other, more formal approaches focus solely on the systems' input-output relations [13], [42], without considering multiple invocations of the agent and its interactions with the environment. Thus, previous methods are not able to provide rigorous guarantees regarding the correctness of multistep robotic systems, and do not take into account sequential decision making -which renders these methods insufficient for detecting various safety and liveness violations.…”
Section: Related Workmentioning
confidence: 99%
“…To perform the training of our robot, we rely on Unity3D, a the popular engine originally designed for game development, that has recently been adopted for robotics simulation [42], [50]. In particular, the built-in physics engine, the powerful 3D rendering algorithm and the time control system (which allows to speed up the simulation by more than 10 times), have made Unity3D a very powerful tool in these contexts [29].…”
Section: Appendix B Technical Specifications Of the Robotmentioning
confidence: 99%
“…Recently, Reinforcement Learning (RL) [1] has been put to significant uses such as robotic table tennis [2], surgical robot planning [3], rapid motion planning in bimanual regrasping for suture needles [4] and Aquatic Navigation [5]. Each of these applications make use of RL as an encouraging substitute to automating manual effort.…”
Section: Introductionmentioning
confidence: 99%