2020
DOI: 10.1109/access.2020.3014791
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning-Based Access Control for Buffer-Aided Relaying Systems With Energy Harvesting

Abstract: This paper considered a buffer-aided relaying system with multiple source-destination pairs and a relay node (RN) with energy harvesting (EH) capability. The RN harvests energy from the ambient environment and uses the harvested energy to forward the sources' information packet to the corresponding destinations. It is assumed that information on the EH and channel gain processes is unavailable. Thus, the model free deep reinforcement learning (DRL) method, specifically the deep Q-learning, is applied to learn … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 30 publications
0
5
0
Order By: Relevance
“…In [26], Sun et al proposed a dynamic resource reservation and DRLbased autonomous virtual resource slicing framework for the next generation radio access network. Additionally, regarding the wireless access control problems, in [27]- [29], the DRL based routing selection approaches have been proposed, and in [30], [31], the DRL based dynamic power allocation approaches have been investigated. In our previous work [32], we also proposed a DRL-based radio access control mechanism for disaster response network, but without the consideration of UE's mobility and buffering capacity.…”
Section: Related Workmentioning
confidence: 99%
“…In [26], Sun et al proposed a dynamic resource reservation and DRLbased autonomous virtual resource slicing framework for the next generation radio access network. Additionally, regarding the wireless access control problems, in [27]- [29], the DRL based routing selection approaches have been proposed, and in [30], [31], the DRL based dynamic power allocation approaches have been investigated. In our previous work [32], we also proposed a DRL-based radio access control mechanism for disaster response network, but without the consideration of UE's mobility and buffering capacity.…”
Section: Related Workmentioning
confidence: 99%
“…In [155], a low-complexity mechanism for relay scheduling in cooperative IoT networks using a stateless RL method -the multiarmed bandit (MAB) -is investigated. The authors utilized the MAB framework to learn relay scheduling using only the acknowledgments (and negative acknowledgments) of packet transmissions.…”
Section: B Reinforcement Learningmentioning
confidence: 99%
“…Thus, AI-based algorithms can avoid the massive iterations and alleviate the computation overhead of conventional methods. Moreover, the RL and DRL techniques can efficiently address the problem of the huge size of solution space [171], [179]. Furthermore, the combinations of ML/DL models with heuristic algorithms or game theory can further enhance efficiency [79], [117], [134], [161], [180] IV.…”
Section: E Summarymentioning
confidence: 99%