2022 International Conference on Big Data, Information and Computer Network (BDICN) 2022
DOI: 10.1109/bdicn55575.2022.00098
|View full text |Cite
|
Sign up to set email alerts
|

An End-to-end Flow Control Method Based on DQN

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 9 publications
0
2
0
Order By: Relevance
“…DDPG uses the experience replay of Deep Q-learning (Gao and Jin, 2022), and adds two target networks, namely, the targetactor network and the target-critic network. The loss function L of the critic network is defined as…”
Section: Td3 Algorithmmentioning
confidence: 99%
“…DDPG uses the experience replay of Deep Q-learning (Gao and Jin, 2022), and adds two target networks, namely, the targetactor network and the target-critic network. The loss function L of the critic network is defined as…”
Section: Td3 Algorithmmentioning
confidence: 99%
“…DRL's foray into the domain of path planning is particularly noteworthy, where it demonstrates remarkable competence in negotiating complex environments and executing challenging tasks with notable alacrity. The Deep Q-Network (DQN) paradigm [9]- [11], an amalgamation of Q-learning and deep learning, epitomizes this synergy by transforming the state-action value function into a construct amenable to neural network interpretation, thereby enabling the calculation of action values pertinent to the current state and facilitating the identification of an optimal navigational strategy.…”
Section: Introductionmentioning
confidence: 99%