2018
DOI: 10.1016/j.neucom.2017.06.066
|View full text |Cite
|
Sign up to set email alerts
|

Concise deep reinforcement learning obstacle avoidance for underactuated unmanned marine vessels

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
77
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 202 publications
(78 citation statements)
references
References 15 publications
1
77
0
Order By: Relevance
“…As described in the article, the approach is model-free, requiring no prior knowledge of the system it is assigned to control. Another example of RL in USV steering is [3] who propose a deep RL approach for obstacle avoidance. As RL approaches, these require a pre-made reward function.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…As described in the article, the approach is model-free, requiring no prior knowledge of the system it is assigned to control. Another example of RL in USV steering is [3] who propose a deep RL approach for obstacle avoidance. As RL approaches, these require a pre-made reward function.…”
Section: Related Workmentioning
confidence: 99%
“…They are able to handle highly nonlinear relationships between their input and output [12,4], an ability which is necessary in order to perform end-to-end steering in USVs. Systems that couple deep ANNs with Reinforcement Learning (RL), have also proved to be able to learn complex tasks [14,11,10,8,3,7,15]. However, RL requires a known reward function to guide its training.…”
Section: Introductionmentioning
confidence: 99%
“…Considering ocean currents effect, a path planning algorithm based on RL is designed [237]. Aiming at obstacle avoidance, a concise deep reinforcement learning obstacle avoidance (CDRLOA) algorithm is proposed to overcome the usability issue caused by complicated control laws in traditional analysis methods [238].…”
Section: Artificial Intelligencementioning
confidence: 99%
“…Equation 4 allows us to obtain a smoother transition between risk-free states (i.e., known states) and risk states (i.e., unknown states). The parameter k has a double effect.…”
Section: The Continuous Risk Functionmentioning
confidence: 99%
“…There are many RL-based approaches to learn robotic tasks, but most of them do not face the problem of avoiding undesirable situations during training [4,13,17,29]. Commonly in these approaches, the agent receives a large negative reward for each visit to an undesirable or dangerous situation, and it must repeatedly experience these (possibly catastrophic) situations to learn how to avoid them.…”
Section: Introductionmentioning
confidence: 99%