2023
DOI: 10.3390/drones7070476
|View full text |Cite
|
Sign up to set email alerts
|

Partially Observable Mean Field Multi-Agent Reinforcement Learning Based on Graph Attention Network for UAV Swarms

Abstract: Multiple unmanned aerial vehicles (Multi-UAV) systems have recently demonstrated significant advantages in some real-world scenarios, but the limited communication range of UAVs poses great challenges to multi-UAV collaborative decision-making. By constructing the multi-UAV cooperation problem as a multi-agent system (MAS), the cooperative decision-making among UAVs can be realized by using multi-agent reinforcement learning (MARL). Following this paradigm, this work focuses on developing partially observable … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(7 citation statements)
references
References 38 publications
0
5
0
Order By: Relevance
“…In previous works, the proof of mean field Q‐learning converges to the Nash Q ‐values has been shown in ref. [22], and the convergence of MFQ with double estimators has been proved by ref. [24].…”
Section: Mfdql‐dtc: Architecture and Methodologymentioning
confidence: 99%
See 2 more Smart Citations
“…In previous works, the proof of mean field Q‐learning converges to the Nash Q ‐values has been shown in ref. [22], and the convergence of MFQ with double estimators has been proved by ref. [24].…”
Section: Mfdql‐dtc: Architecture and Methodologymentioning
confidence: 99%
“…Based on this, the mean field approximation is proposed in ref. [22] to reduce the action space's dimension and to handle the convergence problem in MARL. In this theory, the action is encoded by one‐hot coding for computing convenience.…”
Section: Preliminariesmentioning
confidence: 99%
See 1 more Smart Citation
“…As a result, traditional MARL experiences scaling problems, particularly with exponential agent increments. To address this constraint, current research has focused on addressing scaling issues in MARL by adding mean field theory [172].…”
Section: E Scalabilitymentioning
confidence: 99%
“…The objective of each agent is to maximize its total return R i = ∞ t=0 γ t r t i . In Markov Games settings, agents can cooperate [25,71,72], compete with other agents for resources or rewards [18,73] or even betray the cooperation [74]. Many MARL methods solve Markov Games problems [25,71,[75][76][77][78][79][80][81], among which MADDPG [25], MF-Q [71] and PSRO [75] are representative methods.…”
Section: Markov Gamesmentioning
confidence: 99%