GLOBECOM 2020 - 2020 IEEE Global Communications Conference 2020
DOI: 10.1109/globecom42002.2020.9322182
|View full text |Cite
|
Sign up to set email alerts
|

A Reinforcement Learning Framework for QoS-Driven Radio Resource Scheduler

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(3 citation statements)
references
References 11 publications
0
3
0
Order By: Relevance
“…This will make the size of the NN huge and penalize the convergence time. As in [16] and [17], we use the dynamic architecture shown in Fig. 2…”
Section: Reinforcement Learning Modelmentioning
confidence: 99%
“…This will make the size of the NN huge and penalize the convergence time. As in [16] and [17], we use the dynamic architecture shown in Fig. 2…”
Section: Reinforcement Learning Modelmentioning
confidence: 99%
“…The proposed architecture minimizes the neural network's size, reducing the computational requirements. This flexible architecture was also used in [23]. While their RL agent takes just one action per TTI, ours takes one action in a step, and each episode is the collection of decisions made at each step (the selection of a user at each step) in one TTI.…”
Section: Rl Architecturementioning
confidence: 99%
“…The action changes the state in which a reward can be calculated. The RL state consists of information from both channels and queues, generically denoted as CSI and queue state information (QSI), respectively, similar to [54]. Thus, the scheduler is cross-layer since it considers information from layers other than the PHY, such as the buffer occupancy of active users and the age of the packets.…”
Section: A States Actions and Rewardsmentioning
confidence: 99%