2020
DOI: 10.1609/aaai.v34i05.6214
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Agent Actor-Critic with Hierarchical Graph Attention Network

Abstract: Most previous studies on multi-agent reinforcement learning focus on deriving decentralized and cooperative policies to maximize a common reward and rarely consider the transferability of trained policies to new tasks. This prevents such policies from being applied to more complex multi-agent tasks. To resolve these limitations, we propose a model that conducts both representation learning for multiple agents using hierarchical graph attention network and policy learning using multi-agent actor-critic. The hie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 96 publications
(32 citation statements)
references
References 19 publications
(24 reference statements)
0
32
0
Order By: Relevance
“…As such, we can regard the opponent sample complexity as communication load. To lower the opponent sample complexity, we replace the real opponents with learned opponent models in data simulation, which is analogous to selectively call some (or none) of opponents for useful information to reduce the communication load (or bandwidth) in multi-agent interactions [Ryu et al, 2020].…”
Section: Two Parts Of Sample Complexitymentioning
confidence: 99%
“…As such, we can regard the opponent sample complexity as communication load. To lower the opponent sample complexity, we replace the real opponents with learned opponent models in data simulation, which is analogous to selectively call some (or none) of opponents for useful information to reduce the communication load (or bandwidth) in multi-agent interactions [Ryu et al, 2020].…”
Section: Two Parts Of Sample Complexitymentioning
confidence: 99%
“…To address the limitation, the agent grouping method [12] employs a two-level graph neural network to model the interagent and intergroup relationships effectively. However, it ignores the communication relationship between the agents in the same group.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, deep reinforcement learning (DRL) has shown great potential in many domains, such as games [5,6] and robotics [7,8]. Inspired by the powerful perception and learning ability of DRL, researchers have made continuous attempts to apply DRL to multiagent reinforcement learning (MARL) to promote multiagent cooperative behaviors in environments with many agents [9][10][11][12][13][14][15]. Based on the common paradigm of centralized learning with decentralized execution, some MARL algorithms learn centralized critics for multiple agents and determine the decentralized action.…”
Section: Introductionmentioning
confidence: 99%
“…In RL and MARL, various forms of inductive bias have been used to improve learning. The most straightforward inductive biases entail designing network structures for the critic or policy, such as attention networks [5], graph neural networks [15], and implicit communication structures [14]. However, biases in game information, such as state, reward, and action have also been used in an attempt to boost training.…”
Section: Biases In Rl and Marlmentioning
confidence: 99%