2020
DOI: 10.1109/access.2020.3016951
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Robot Flocking Control Based on Deep Reinforcement Learning

Abstract: In this paper, we apply deep reinforcement learning (DRL) to solve the flocking control problem of multi-robot systems in complex environments with dynamic obstacles. Starting from the traditional flocking model, we propose a DRL framework for implementing multi-robot flocking control, eliminating the tedious work of modeling and control designing. We adopt the multi-agent deep deterministic policy gradient (MADDPG) [1] algorithm, which additionally uses the information of multiple robots in the learning proce… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
28
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 58 publications
(29 citation statements)
references
References 21 publications
0
28
0
1
Order By: Relevance
“…The flocking behavior presents interesting characteristics that make it of high interest for the design of artificial systems, particularly in problems of localization, search and rescue. This type of behavior has been observed in birds, is similar to schooling fish and swarming insects, and is characterized by a joint movement of the group without central coordination [12], [13]. The first basic rules of this dynamic were established in 1987 as alignment, cohesion, and separation [14].…”
Section: Introductionmentioning
confidence: 86%
“…The flocking behavior presents interesting characteristics that make it of high interest for the design of artificial systems, particularly in problems of localization, search and rescue. This type of behavior has been observed in birds, is similar to schooling fish and swarming insects, and is characterized by a joint movement of the group without central coordination [12], [13]. The first basic rules of this dynamic were established in 1987 as alignment, cohesion, and separation [14].…”
Section: Introductionmentioning
confidence: 86%
“…MADDPG has been used in many applications such as Wang et al [33] that proposed a data-driven multiagent power grid control scheme using MADDPG for the large-scale energy system with more control options and operating conditions. Zhu et al [34] applied MADDPG to solve the flocking control problem of multi-robot systems in complex environments with dynamic obstacles. Lei et al [35] introduced edge computing between terminals and the cloud using MADDPG to address the drawbacks of the traditional power cloud paradigm.…”
Section: Multiagent Reinforcement Learningmentioning
confidence: 99%
“…In [6], although the authors realized their work based on LiDAR and an odometer, they only considered the -greedy policy of the DQN with different parameters to update the neural network. In [7], the method was implemented based on virtual robots. Namely, rather than using a simulated model, they directly assumed the effectiveness of the simulation properties of a virtual robot, e.g., the gyration radius and mass, the maximum speed and the maximum acceleration, and so on.…”
Section: Introductionmentioning
confidence: 99%