2022
DOI: 10.48550/arxiv.2205.14205
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ALMA: Hierarchical Learning for Composite Multi-Agent Tasks

Abstract: Despite significant progress on multi-agent reinforcement learning (MARL) in recent years, coordination in complex domains remains a challenge. Work in MARL often focuses on solving tasks where agents interact with all other agents and entities in the environment; however, we observe that real-world tasks are often composed of several isolated instances of local agent interactions (subtasks), and each agent can meaningfully focus on one subtask to the exclusion of all else in the environment. In these composit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 14 publications
0
3
0
Order By: Relevance
“…The task allocation problem is solved by search algorithms, and task execution is solved by coordinated reinforcement learning. The work [4] proposes a hierarchical method to learn both high-level subtask allocation policy and low-level agent policies. The high-level policy considers the states of all subtasks as input to estimate values for different allocations.…”
Section: Reinforcement Learning Methods For Allocation Problemsmentioning
confidence: 99%
See 2 more Smart Citations
“…The task allocation problem is solved by search algorithms, and task execution is solved by coordinated reinforcement learning. The work [4] proposes a hierarchical method to learn both high-level subtask allocation policy and low-level agent policies. The high-level policy considers the states of all subtasks as input to estimate values for different allocations.…”
Section: Reinforcement Learning Methods For Allocation Problemsmentioning
confidence: 99%
“…In this section, we examine the performance of the proposed method on two challenging environments, i.e., Google Research Football (GRF) [22] and SAVETHECITY [4]. We aim to answer three questions in the following experiments:…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation