2019
DOI: 10.3390/s19183837
|View full text |Cite
|
Sign up to set email alerts
|

Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning

Abstract: In this paper, we propose a novel Deep Reinforcement Learning (DRL) algorithm which can navigate non-holonomic robots with continuous control in an unknown dynamic environment with moving obstacles. We call the approach MK-A3C (Memory and Knowledge-based Asynchronous Advantage Actor-Critic) for short. As its first component, MK-A3C builds a GRU-based memory neural network to enhance the robot’s capability for temporal reasoning. Robots without it tend to suffer from a lack of rationality in face of incomplete … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
38
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 57 publications
(38 citation statements)
references
References 21 publications
0
38
0
Order By: Relevance
“…As a variant of LSTM, the Gated Recurrent Unit (GRU) is easier to train. Zeng et al [80] built a GRU-based memory neural network and proved that it could improve performance in complex navigation tasks.…”
Section: Solutionmentioning
confidence: 99%
“…As a variant of LSTM, the Gated Recurrent Unit (GRU) is easier to train. Zeng et al [80] built a GRU-based memory neural network and proved that it could improve performance in complex navigation tasks.…”
Section: Solutionmentioning
confidence: 99%
“…Zeng et al [113] proposed an Actor-Critic method (MK-A3C) based on asynchronous advantages of memory and knowledge for the navigation problem of nonholonomic robots with continuous control. MK-A3C constructs a memory neural network based on GRU to enhance the temporal reasoning ability of the robot.At the same time, the robot is given a certain memory ability.…”
Section: (A) (B) Figure 17 Navigating In Complex Environments Based On A3cmentioning
confidence: 99%
“…There are several path planning algorithms developed for robotics systems, such as grid-based search algorithms (assumes every point/object is covered in a grid configuration [44,45]), interval-based search algorithms (generates paving to cover an entire configuration space instead of grid [44]), geometric algorithms (find safe path from the start to goal initially [46]), reward-based algorithms (a robot tries to take a path, and it is rewarded positively if successful and negatively if otherwise [47]), artificial potential fields algorithms (robot is modeled to be attracted to positive path and repelled by obstacles [48]), and sampling-based algorithms (path is found from the roadmap spaces of the configuration space). Each of the algorithms has potential use, and some are just classic methods like grid-based algorithms [49].…”
Section: Agricultural Robot Path Planningmentioning
confidence: 99%