2020
DOI: 10.1109/tvt.2020.3020400
|View full text |Cite
|
Sign up to set email alerts
|

Joint Optimization of Handover Control and Power Allocation Based on Multi-Agent Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
45
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 119 publications
(45 citation statements)
references
References 24 publications
0
45
0
Order By: Relevance
“…The proposed model uses the received signal measurements to reduce the number of unnecessary HOs and predict the user location while ensuring that the throughput of the network is maintained. A joint optimization framework for minimizing HO frequency and maximizing user throughput was proposed in [178]. The HO and power allocation problem was modeled as a cooperative multi-agent task, after which a MARL framework using proximal policy optimization (PPO) was developed.…”
Section: ) Wireless Data Aided Handover Optimizationmentioning
confidence: 99%
“…The proposed model uses the received signal measurements to reduce the number of unnecessary HOs and predict the user location while ensuring that the throughput of the network is maintained. A joint optimization framework for minimizing HO frequency and maximizing user throughput was proposed in [178]. The HO and power allocation problem was modeled as a cooperative multi-agent task, after which a MARL framework using proximal policy optimization (PPO) was developed.…”
Section: ) Wireless Data Aided Handover Optimizationmentioning
confidence: 99%
“…A handover penalty is added to the reward function to reduce the redundant handovers. Similarly, in [15] the authors propose a proximal policy optimisation method based on a multi-agent reinforcement learning algorithm for handover control and power allocation in HetNets. The handover in [15] is triggered based on a hybrid state vector, i.e., the UE current serving cell, UE's signal measurement, and the number of UEs served by each BS.…”
Section: Intelligent Mobility Management Controlmentioning
confidence: 99%
“…Similarly, in [15] the authors propose a proximal policy optimisation method based on a multi-agent reinforcement learning algorithm for handover control and power allocation in HetNets. The handover in [15] is triggered based on a hybrid state vector, i.e., the UE current serving cell, UE's signal measurement, and the number of UEs served by each BS. Zhou et al in [16] propose two mobility management controls based on an online-learning algorithm, i.e., upper-confidence bound policies.…”
Section: Intelligent Mobility Management Controlmentioning
confidence: 99%
“…Motivated by the deep reinforcement learning (DRL) approach in solving the dynamic problem [13], DRL has been applied into cache policies to improve the cache performance of dynamic cache scenarios. In [14], a DRL approach was proposed to reduce the transmission cost by jointly considering proactive cache and content recommendations.…”
Section: Introductionmentioning
confidence: 99%