2020
DOI: 10.1016/j.ifacol.2020.12.2021
|View full text |Cite
|
Sign up to set email alerts
|

A Multi-Agent Off-Policy Actor-Critic Algorithm for Distributed Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 35 publications
(40 citation statements)
references
References 10 publications
0
40
0
Order By: Relevance
“…Further work could be devoted to the weak convergence analysis of alternative multi-agent temporal difference schemes, including the emphatic temporal difference algorithm [38,44] and actor-critic algorithms [36,37]. Also, the proposed schemes could be extended to the cases of nonlinear value function approximations (such as using deep neural networks [27]).…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Further work could be devoted to the weak convergence analysis of alternative multi-agent temporal difference schemes, including the emphatic temporal difference algorithm [38,44] and actor-critic algorithms [36,37]. Also, the proposed schemes could be extended to the cases of nonlinear value function approximations (such as using deep neural networks [27]).…”
Section: Discussionmentioning
confidence: 99%
“…The proposed distributed multi-agent algorithms can be considered as: 1) a tool for organizing coordinated actions of multiple agents contributing to the value function estimation and 2) a parallelization tool, allowing faster convergence, useful particularly in the problems with large dimensions. Notice that the in the first case the proposed algorithms can become a part of multi agent actor-critic schemes (see, e.g., [36,37]).…”
Section: Inter-agent Communications and Network Designmentioning
confidence: 99%
See 1 more Smart Citation
“…Finite-sample analyses on multi-agent reinforcement learning are also extended to fitted Q-iterations [Zhang et al, 2021]. For multi-agent actor-critic methods, the best available result is the proof of the asymptotic convergence [Suttle et al, 2020].…”
Section: Analysis Of Multi-agent Reinforcement Learningmentioning
confidence: 99%
“…In the decentralized setting, a few works have obtained the almost sure convergence results of AC [17,21,22,35,36], but the finite-time convergence rate and complexity are still unexplored. To the best of our knowledge, there is no formally developed decentralized NAC algorithm.…”
Section: Related Workmentioning
confidence: 99%