2019
DOI: 10.48550/arxiv.1902.06740
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Leveraging Communication Topologies Between Learning Agents in Deep Reinforcement Learning

Abstract: A common technique to improve speed and robustness of learning in deep reinforcement learning (DRL) and many other machine learning algorithms is to run multiple learning agents in parallel. A neglected component in the development of these algorithms has been how best to arrange the learning agents involved to better facilitate distributed search. Here we draw upon results from the networked optimization and collective intelligence literatures suggesting that arranging learning agents in less than fully conne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(6 citation statements)
references
References 22 publications
0
6
0
Order By: Relevance
“…We now establish policy improvement theorem of stochastic TAPE, and prove a theorem for the cooperation improvement from the perspective of exploring the parameter space, which is a common motivation in RL research (Schulman, Chen, and Abbeel 2017;Haarnoja et al 2018;Zhang et al 2021;Adjodah et al 2019). We assume the policy to have tabular expressions.…”
Section: Theoretical Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…We now establish policy improvement theorem of stochastic TAPE, and prove a theorem for the cooperation improvement from the perspective of exploring the parameter space, which is a common motivation in RL research (Schulman, Chen, and Abbeel 2017;Haarnoja et al 2018;Zhang et al 2021;Adjodah et al 2019). We assume the policy to have tabular expressions.…”
Section: Theoretical Resultsmentioning
confidence: 99%
“…As the global Q value is determined by the centralized critic for all agents, sub-optimal actions of one agent will easily influence all others. Topology in Reinforcement Learning Adjodah et al (Adjodah et al 2019) discuss the communication topology issue in parallel-running RL algorithms such as A3C (Mnih et al 2016). Results show that the centralized learner implicitly yields a fully-connected communication topology among parallel workers, which will harm their performance.…”
Section: Related Workmentioning
confidence: 99%
“…• Fast aggregation via over-the-air computation [21], [84], [85], [86] • Aggregation frequency control with limited bandwidth and computation resources [87], [88], [89] • Data reshuffling via index coding and pliable index coding for improving training performance [90], [91], [92] • Straggler mitigation via coded computing [93], [94], [95], [96], [97], [98], [99], [100], [101] • Training in decentralized system mode [102], [103], [104], [105], [106], [107], [108], [109], [110], [111], [112] Model Partition Based Edge Training Systems…”
Section: Data Partition Based Edge Training Systemsmentioning
confidence: 99%
“…There have been several works demonstrating that some carefully designed topologies of networks achieve better performance than the fully connected network. It has been empirically observed in [107] that using an alternative network topology between devices can lead to improved learning performance in several deep reinforcement learning tasks compared with the standard fully-connected communication topology. Specifically, it was observed in [107] that the Erdos-Renyi graph topology with 1000 devices can compete with the standard fully-connected topology with 3000 devices, which shows that the machine learning performance can be more efficient if the topology is carefully designed.…”
Section: Awgn Receive Beamformermentioning
confidence: 99%
See 1 more Smart Citation