2022
DOI: 10.1109/tnnls.2020.3029711
|View full text |Cite
|
Sign up to set email alerts
|

Deep Echo State Q-Network (DEQN) and Its Application in Dynamic Spectrum Sharing for 5G and Beyond

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 38 publications
(14 citation statements)
references
References 16 publications
0
14
0
Order By: Relevance
“…Zhang et al and some people say that digital technology innovation represented by big data and artificial intelligence has brought subversive changes and innovations to all walks of life. As a key stage in the empowerment of the digital economy, smart cities have once again become the focus of attention [4]. Cloud computing is shown in Figure 2.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Zhang et al and some people say that digital technology innovation represented by big data and artificial intelligence has brought subversive changes and innovations to all walks of life. As a key stage in the empowerment of the digital economy, smart cities have once again become the focus of attention [4]. Cloud computing is shown in Figure 2.…”
Section: Literature Reviewmentioning
confidence: 99%
“…This study uses deep echo state Q-network [23] in which the agent stores the observation o and the reservoir output x including the memory of the context of the observation. Fig.…”
Section: Deep Echo State Q-networkmentioning
confidence: 99%
“…Chang proposed deep echo state Q-network in which the system stores the output of the ESN in replay memory to avoid the issues of large computational cost and vanishing/exploding gradients problem [23]. The basic idea of this approach is to input the observation from the environment to the reservoir and store the observations and the reservoir output containing the spatio-temporally expanded memory of the observations into the replay memory.…”
Section: Introductionmentioning
confidence: 99%
“…The outline of FL in MARL can be found in Figure 1. We select Signal-to-Interference-plus-Noise Ratio (SINR) as our quality measurement [9] which characterises all the factors to SUs network, such as background noise, the transmission power, the interference between simultaneously transmission pairs. In addition, we choose channel capacity as the reward function.…”
Section: Configuration Of Marl Enabled Flmentioning
confidence: 99%
“…FedProx modifies the FedAvg approach somewhat by allowing for partial work to be distributed among devices based on the underlying system restrictions and then using the proximal term to securely include the partial work. Furthermore, when applying the FL framework in RL, the widely used Q-learning or DQN (discrete actions) in solving DSS problems [9], [15] requires a solid theoretic proof when making convergence analysis based on FedAVG or FedProx. Therefore, policy gradient as an alternative approach is introduced in this article to update the shared model by averaging aggregated local policies.…”
Section: Asynchronous Fl Optimizationmentioning
confidence: 99%