2022
DOI: 10.1109/jiot.2021.3091142
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning for Energy-Efficient Computation Offloading in Mobile-Edge Computing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
67
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8
1
1

Relationship

1
9

Authors

Journals

citations
Cited by 195 publications
(67 citation statements)
references
References 42 publications
0
67
0
Order By: Relevance
“…Compared with outdoor air quality, indoor air quality (IAQ) is more significant because people spend nearly 80% of their time staying in indoor environments (homes, schools, offices, and so on) [3,4]. Therefore, there is an urgent need to conduct effective IAQ monitoring and control for avoiding respiratory illness and protecting people's health [5,6].…”
Section: Introductionmentioning
confidence: 99%
“…Compared with outdoor air quality, indoor air quality (IAQ) is more significant because people spend nearly 80% of their time staying in indoor environments (homes, schools, offices, and so on) [3,4]. Therefore, there is an urgent need to conduct effective IAQ monitoring and control for avoiding respiratory illness and protecting people's health [5,6].…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, hybrid channel access with nonorthogonal multiple access (NOMA) and orthogonal multiple access (OMA) users was considered in [128] to achieve the optimal policy of partial offloading decisions and channel resource allocation using actor-critic DQN. In the time-varying MEC system with multiple users accessing an individual MEC server, to minimize the total energy consumption, the offloading strategy was proposed in [129] by considering the heterogeneous resource requirements and delay constraints in communication and computation. In this regard, Q-learning and its low-complexity version, i.e., double DQN, are adopted to achieve the optimal strategy.…”
Section: ) Dynamic Spectrum Accessmentioning
confidence: 99%
“…The Q-learning, DQN, and Dueling DQN are all valuebased deep reinforcement models that can formulate the dynamic offloading problem as a Markov decision process, using DQN to approximate the objective value function to obtain an optimized offloading policy, even for multi-user mobile edge computing networks, DQN-based algorithms have advantages over traditional algorithms [30][31][32]. DQN may suffer from overestimation due to choosing the maximum action value each time in order to avoid this situation, Double DQN are generated, in [33], delay constraints and uncertain resource requirements of heterogeneous computing tasks, and in order to avoid dimensional disasters and overestimation using DDQN-based algorithms, which eventually proved the effectiveness of the algorithm. It is also possible to change the network structure of DQN to produce Dueling DQN, [34] firstly preprocesses the data and then uses Dueling DQN for optimization of the objective function, its convergence speed is faster than DQN.…”
Section: Related Workmentioning
confidence: 99%