2017
DOI: 10.1016/j.neucom.2016.08.108
|View full text |Cite
|
Sign up to set email alerts
|

Path planning of multi-agent systems in unknown environment with neural kernel smoothing and reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 57 publications
(10 citation statements)
references
References 29 publications
0
10
0
Order By: Relevance
“…Table 1 summarizes the DRL multi-robot path planning methods and the advantages and limitations of each method. From the information in Table 1, it can be summarized that shared parameter type algorithms such as MADDPG and ME-MADDPG can be used in dynamic and complex environments [1][2][3][4] ; decentralized architectures such as DQN and DDQN can be considered in stable environments [5][6][7] ; large robotic systems facing a large number of dynamic obstacles can be considered using algorithms such as A2C, A3C and TDueling [8][9][10][11] . Validity validated on only a few teams of agents.…”
Section: Drl Multi-robot Path Planning Methodsmentioning
confidence: 99%
“…Table 1 summarizes the DRL multi-robot path planning methods and the advantages and limitations of each method. From the information in Table 1, it can be summarized that shared parameter type algorithms such as MADDPG and ME-MADDPG can be used in dynamic and complex environments [1][2][3][4] ; decentralized architectures such as DQN and DDQN can be considered in stable environments [5][6][7] ; large robotic systems facing a large number of dynamic obstacles can be considered using algorithms such as A2C, A3C and TDueling [8][9][10][11] . Validity validated on only a few teams of agents.…”
Section: Drl Multi-robot Path Planning Methodsmentioning
confidence: 99%
“…Cruz and Yu [35] proposed a method that combines kernel smoothing and the DRL of the WoLF-Policy Hill Climbing algorithm to solve the difficulty of traditional reinforcement learning in path planning in an unfamiliar environment. Without prior knowledge, the discrete action space was used to approximate the state of MARL by kernel smoothing, thereby reducing the state space in the Q table.…”
Section: Path Planning Approach Based On Rlmentioning
confidence: 99%
“…The idea of utilizing Q-learning with the obstacle aware to generate the shortest track from the source to destination in a grid-based divided sub-region was been proposed in Aleksandr et al [ 18 ], Amit et al [ 19 ] and Soong et al [ 20 ]. David et al amplifies this strategy to different robot specialists [ 21 ]. Yuan et al [ 22 ] utilized the RNN gated recurrent unit (GRU) framework to plan an ideal way from the source to the destination straightforwardly.…”
Section: Introductionmentioning
confidence: 99%