2021
DOI: 10.1109/jiot.2020.3014926
|View full text |Cite
|
Sign up to set email alerts
|

Deep-Reinforcement-Learning-Based Proportional Fair Scheduling Control Scheme for Underlay D2D Communication

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
42
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 70 publications
(42 citation statements)
references
References 32 publications
0
42
0
Order By: Relevance
“…We build the two deep neural networks, consisting of three fully-connected layers with 64, 64 and 32 neurons. The computational complexity of the proposed ADQN algorithm for training is W T × W P as the number of iterations of loops in [39], [40], the priori information does not introduce extra complexity. Moreover, after training the algorithm, the complexity of ADQN for making decisions depends on the structure of the neural network.…”
Section: Asynchronous Dqn-based Schemementioning
confidence: 99%
“…We build the two deep neural networks, consisting of three fully-connected layers with 64, 64 and 32 neurons. The computational complexity of the proposed ADQN algorithm for training is W T × W P as the number of iterations of loops in [39], [40], the priori information does not introduce extra complexity. Moreover, after training the algorithm, the complexity of ADQN for making decisions depends on the structure of the neural network.…”
Section: Asynchronous Dqn-based Schemementioning
confidence: 99%
“…4, and the specific implementation process of DDQNPI-RP scheme is shown in Algorithm 2. The computational complexity of the proposed algorithm with/without the priori information is V (M + W ) as the number of iterations of loops in [39], [40], exploring the priori information does not introduce extra complexity.…”
Section: Learning Algorithm With the Priori Informationmentioning
confidence: 99%
“…I. Budhiraja et al [39] propose communication resource and power allocation control based on deep reinforcement learning. Communication resource and transmission power are allocated to each D2D device so that minimum acceptable Signal-to-Interference and Noise Ratio (SINR) is kept for D2D devices and cellular users while mitigating interference among them.…”
Section: Radio Interference Avoidance In Wireless Networkmentioning
confidence: 99%