2022
DOI: 10.1016/j.neucom.2021.09.071
|View full text |Cite
|
Sign up to set email alerts
|

Path planning and dynamic collision avoidance algorithm under COLREGs via deep reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 53 publications
(20 citation statements)
references
References 19 publications
0
20
0
Order By: Relevance
“…The purpose of path planning is to reach the target point, but previous research has mainly focused on reaching a single goal in dynamic or unpredictable environments. [3], [11], [16], [17], [42], [68]. The goal should be pre-defined by the user and the agent who completed learning can reach only the defined goal.…”
Section: A Advantage Of the Proposed Methodsmentioning
confidence: 99%
“…The purpose of path planning is to reach the target point, but previous research has mainly focused on reaching a single goal in dynamic or unpredictable environments. [3], [11], [16], [17], [42], [68]. The goal should be pre-defined by the user and the agent who completed learning can reach only the defined goal.…”
Section: A Advantage Of the Proposed Methodsmentioning
confidence: 99%
“…If non-uniform sampling is adopted, the learning rate al pha must be adjusted according to the sampling probability. Set the learning rate as Equation (10).…”
Section: The Improved Dqnmentioning
confidence: 99%
“…It was first proposed by Sutton, whose essence is to learn from the interaction. In recent years, there has been a boom in theoretical research and technical implementation of reinforcement learning, especially in the Deep Q-Network (DQN) algorithm [10,11] proposed by the Google team. In the shipping industry, reinforcement learning is widely used in collision avoidance.…”
Section: Introductionmentioning
confidence: 99%
“…A well-suited machine learning approach to solve motion planning tasks in uncertain environments is RL [5], [6], [34], [35]. The regarded scenario is usually on the open sea with other non-reactive dynamic obstacles [6], [35] and static obstacles [5], [34].…”
Section: Related Workmentioning
confidence: 99%