2020
DOI: 10.1109/access.2020.2985679
|View full text |Cite
|
Sign up to set email alerts
|

Joint Demand Forecasting and DQN-Based Control for Energy-Aware Mobile Traffic Offloading

Abstract: With the explosive growth in demand for mobile traffic, one of the promising solutions is to offload cellular traffic to small base stations for better system efficiency. Due to increasing system complexity, network operators are facing severe challenges and looking for machine learning-based solutions. In this work, we propose an energy-aware mobile traffic offloading scheme in the heterogeneous network jointly apply deep Q network (DQN) decision making and advanced traffic demand forecasting. The base statio… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(11 citation statements)
references
References 35 publications
0
11
0
Order By: Relevance
“…Also, some important parameters are listed in Table 2. Moreover, In (28 further improve the learning efficiency. Furthermore, when dividing the computational resources of each FAP, in order to avoid missing the optimal solution, we set that each resource block is 0.1 Figures 5 and 6 display the learning curves representing the accumulated reward obtained by the agent in each training epoch of the DDQN algorithm and the DQN algorithm, respectively.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Also, some important parameters are listed in Table 2. Moreover, In (28 further improve the learning efficiency. Furthermore, when dividing the computational resources of each FAP, in order to avoid missing the optimal solution, we set that each resource block is 0.1 Figures 5 and 6 display the learning curves representing the accumulated reward obtained by the agent in each training epoch of the DDQN algorithm and the DQN algorithm, respectively.…”
Section: Resultsmentioning
confidence: 99%
“…Environment & State: The environment in RL is defined as the set of all possible states, and the essence of RL is to perform actions to cause the state transfer [28]. Therefore, we set a matrix S as the state, which has the same shape as the matrix P , and the value of s ij in the matrix S should only be 0 or 1, s ij = 1 represents the agent selects FAP i (or DCN i) for UE j, otherwise s ij = 0 .…”
Section: Markov Decision Processmentioning
confidence: 99%
“…Deep Q network (DQN) was introduced in Huang et al [19] for forecasting traffic under high traffic loads. The designed network was not capable of reducing time and memory consumptions.…”
Section: Related Researchmentioning
confidence: 99%
“…Environment & State: The environment in RL is defined as the set of all possible states, and the essence of RL is to perform actions to cause the state transfer [24]. Therefore, we set a matrix S as the state, which has the same shape as the matrix P , and the value of s ij in the matrix S should only be 0 or 1, s ij = 1 represents the agent selects FAP i (or DCN i) for UE j, otherwise s ij = 0.…”
Section: ) Markov Decision Processmentioning
confidence: 99%
“…Where ω and ϕ are the network parameters for V (s t ) and A(s t , a t ), respectively. Specifically, V (s t ) stands for the excepted accumulated reward at the state s t , and A(s t , a t ) indicates the degree of superiority of action a t over the average level in state s t presented as formula (24) and (25).…”
Section: ) Ddqn Algorithmmentioning
confidence: 99%