2023
DOI: 10.48550/arxiv.2302.10418
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MAC-PO: Multi-Agent Experience Replay via Collective Priority Optimization

Abstract: Experience replay is crucial for off-policy reinforcement learning (RL) methods. By remembering and reusing the experiences from past different policies, experience replay significantly improves the training efficiency and stability of RL algorithms. Many decisionmaking problems in practice naturally involve multiple agents and require multi-agent reinforcement learning (MARL) under centralized training decentralized execution paradigm. Nevertheless, existing MARL algorithms often adopt standard experience rep… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 41 publications
0
1
0
Order By: Relevance
“…Multi-party collaborative learning (i.e, distributed and federated learning) has wide applications in data mining and machine learning problems [16,28,31,35,43,59,62]. Multi-party collaborative learning in this paper has a more general definition that does not rely on the IID assumption of data to guarantee the convergence analysis.…”
Section: Serverless Multi-party Collaborative Learningmentioning
confidence: 99%
“…Multi-party collaborative learning (i.e, distributed and federated learning) has wide applications in data mining and machine learning problems [16,28,31,35,43,59,62]. Multi-party collaborative learning in this paper has a more general definition that does not rely on the IID assumption of data to guarantee the convergence analysis.…”
Section: Serverless Multi-party Collaborative Learningmentioning
confidence: 99%