2022 30th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP) 2022
DOI: 10.1109/pdp55904.2022.00020
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating Distributed Deep Reinforcement Learning by In-Network Experience Sampling

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 2 publications
0
5
0
Order By: Relevance
“…Overall, our proposed approach is classified as an improvement method for sending more useful experiences to improve the transfer efficiency in the edge-cloud settings, but the prior works mentioned above do not focus on edge-cloud implementation where actors and learner's communication is long-haul as assumed in this paper. Although our previous work [12] also focused on such an edge-cloud environment as mentioned in Section 1, this paper is quite different from the previous work since the previous work proposed an acceleration method of communication by DPDK.…”
Section: Improvement Methods Of Transfer Efficiency Of Distributed Re...mentioning
confidence: 92%
See 3 more Smart Citations
“…Overall, our proposed approach is classified as an improvement method for sending more useful experiences to improve the transfer efficiency in the edge-cloud settings, but the prior works mentioned above do not focus on edge-cloud implementation where actors and learner's communication is long-haul as assumed in this paper. Although our previous work [12] also focused on such an edge-cloud environment as mentioned in Section 1, this paper is quite different from the previous work since the previous work proposed an acceleration method of communication by DPDK.…”
Section: Improvement Methods Of Transfer Efficiency Of Distributed Re...mentioning
confidence: 92%
“…In this case, gradient aggregation between computers becomes a performance bottleneck; thus, they propose to accelerate this process by performing the gradient aggregation within network switches. Our past work [12] also proposes an acceleration method of communication between actors and learner in Ape-X style architecture using DPDK and F-Stack.…”
Section: Improvement Methods Of Transfer Efficiency Of Distributed Re...mentioning
confidence: 99%
See 2 more Smart Citations
“…The DQN is a reinforcement learning algorithm that combines deep neural networks with the Q-learning algorithm. The utilization of a profound neural network is employed for the purpose of forecasting the Q-value function, a theoretical construct that has been investigated [6]. The learning process of this network involves the minimization of the discrepancy between the anticipated and intended Q-values.…”
Section: Dqnmentioning
confidence: 99%