2020
DOI: 10.1016/j.comcom.2020.05.037
|View full text |Cite
|
Sign up to set email alerts
|

“DRL + FL”: An intelligent resource allocation model based on deep reinforcement learning for Mobile Edge Computing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
19
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 40 publications
(19 citation statements)
references
References 16 publications
0
19
0
Order By: Relevance
“…[14] designed three algorithms, including heuristic search, the reformulation linearization technique, and semi-definite relaxation, to jointly minimize latency and offloading failure probability. However, conventional methods face enormous challenges in responding to long-term benefits and a complex MEC environment [15]. The above methods that ensure system optimization at a specific status are not suitable for fast fading channels, as the dynamics of tasks and system environments are not considered.…”
Section: Related Workmentioning
confidence: 99%
“…[14] designed three algorithms, including heuristic search, the reformulation linearization technique, and semi-definite relaxation, to jointly minimize latency and offloading failure probability. However, conventional methods face enormous challenges in responding to long-term benefits and a complex MEC environment [15]. The above methods that ensure system optimization at a specific status are not suitable for fast fading channels, as the dynamics of tasks and system environments are not considered.…”
Section: Related Workmentioning
confidence: 99%
“…Category Objectives QEEC [29] QL Response time, CPU utilization MDP_DT [164] QL Cost AGH+QL [114] QL Energy consumption, QoS MRLCO [26] MRL Network traffic, service latency DeepRM_Plus [46] DRL Turnaround time, cycling time DERP [160] DRL Automatic elasticity DPM [62] DRL Task latency, energy consumption DQST [17] DQL Makespan, load balancing MDRL [48] DDQN Energy consumption, response time RLTS [49] DDQN Makespan DRL-Cloud [115] DDQN Energy consumption, cost ADRL [13] DDQN Resource Utilization, response time IDRQN [60] DDQN Energy consumption, service latency MADRL [50] DDQN Computation delay, channel utilization DDQN [166] DDQN Service latency, system rewards DRL+FL [116] DDQN Energy consumption, load balancing AGH+QL [114], a novel revised Q-learning-based model, takes hash codes as input states with a reduced size of state space. DQST [17], deep Q-learning task scheduling, uses fully connected network to calculate the Q-values which can express the mapper of action decision.…”
Section: Algorithmmentioning
confidence: 99%
“…Actor network with two layer fully connected network is a mapper from state to action, and critic network with two fully connected network hidden layers and an output layer with one node is a mapper from state and action to Q-value. DRL+FL [116], based on DDQN, uses Federal Learning to accelerate the training of DRL agents. MDP_DT [164], a novel full-model based RL for elastic resource management, employs adaptive state space partitioning.…”
Section: Algorithmmentioning
confidence: 99%
See 2 more Smart Citations