2021
DOI: 10.1109/access.2021.3067662
|View full text |Cite
|
Sign up to set email alerts
|

Wireless Access Control in Edge-Aided Disaster Response: A Deep Reinforcement Learning-Based Approach

Abstract: The communication infrastructure is most likely to be damaged after a major disaster occurred, which would lead to further chaos in the disaster stricken area. Modern rescue activities heavily rely on the wireless communications, such as safety status report, disrupted area monitoring, evacuation instruction, rescue coordination, etc. Large amount of data generated from victims, sensors and responders must be delivered and processed in a fast and reliable way, even when the normal communication infrastructure … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 18 publications
(4 citation statements)
references
References 30 publications
0
4
0
Order By: Relevance
“…In order to maximize the expected achievable rate of UE in ultra-dense networks, the authors [ 51 ] developed a matching game algorithm, where mobility-aware user association was considered by minimizing the handovers number. The authors [ 52 ] deployed the DRL algorithm to estimate the transmit timing, routing as well as power allocation for UEs from MDRU deployed in disaster areas where UE mobility, channel states, and energy harvesting were considered.…”
Section: Related Workmentioning
confidence: 99%
“…In order to maximize the expected achievable rate of UE in ultra-dense networks, the authors [ 51 ] developed a matching game algorithm, where mobility-aware user association was considered by minimizing the handovers number. The authors [ 52 ] deployed the DRL algorithm to estimate the transmit timing, routing as well as power allocation for UEs from MDRU deployed in disaster areas where UE mobility, channel states, and energy harvesting were considered.…”
Section: Related Workmentioning
confidence: 99%
“…38 It describes a sequence of exploration–exploitation decision-making processes. 39,40 The MAB model is mainly composed of decision makers, arms, and rewards. 41 In each round, the decision maker selects an arm, and the selected arm will generate a reward.…”
Section: Delay-reliability-aware Mqtt Qos Level Selection In Elotmentioning
confidence: 99%
“…The proposed scheme is based on a typical reinforcement learning algorithm, i.e., Qlearning [25,26]. Reinforcement learning is one of the three basic machine learning paradigms, by which the agent learns the optimal behavior through repeated interactions with the environment in discrete time steps.…”
Section: Proposed Reinforcement Learning-basedmentioning
confidence: 99%