2019
DOI: 10.48550/arxiv.1906.05015
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Reinforcement Learning for Unmanned Aerial Vehicle-Assisted Vehicular Networks

Abstract: Unmanned aerial vehicles (UAVs) are envisioned to complement the 5G communication infrastructure in future smart cities. Hot spots easily appear in road intersections, where effective communication among vehicles is challenging. UAVs may serve as relays with the advantages of low price, easy deployment, line-of-sight links, and flexible mobility. In this paper, we study a UAV-assisted vehicular network where the UAV jointly adjusts its transmission control (power and channel) and 3D flight to maximize the tota… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 35 publications
0
5
0
Order By: Relevance
“…In [155], to jointly design a 3D trajectory for a network of UAVs while at the same time maximizing its throughput, a markov decision process (MDP) problem was formulated. The authors used a deep deterministic policy gradient (DDPG) algorithm to deal with this problem.…”
Section: ) Joint Communication and Fight Control Aspectsmentioning
confidence: 99%
“…In [155], to jointly design a 3D trajectory for a network of UAVs while at the same time maximizing its throughput, a markov decision process (MDP) problem was formulated. The authors used a deep deterministic policy gradient (DDPG) algorithm to deal with this problem.…”
Section: ) Joint Communication and Fight Control Aspectsmentioning
confidence: 99%
“…In [110], to jointly design a 3D trajectory for a network of UAVs while at the same time maximizing its throughput, a markov decision process (MDP) problem was formulated. The authors used deep deterministic policy gradient (DDPG) algorithm to deal with this problem.…”
Section: ) Intelligent Trajectory and Placement Designmentioning
confidence: 99%
“…The aim in [15] was to maximize the number of serving users by proposing a double Q-learning based algorithm. The authors in [16] considered a vehicular control network and utilized an UAV to improve their network performance. They utilized a deep learning algorithm based on Deep Deterministic Policy Gradient (DDPG) for resource allocation and UAV trajectory design.…”
Section: A Motivationmentioning
confidence: 99%
“…UAV u sends signal to user k on subcarrier l through a channel with the channel power gain hl uk (t), which consists of line of sight (LoS) and non-line of sight (NLoS) links as follows [16]:…”
Section: A Communication Model 1) Uav To User Communication (Dl Acces...mentioning
confidence: 99%