GLOBECOM 2020 - 2020 IEEE Global Communications Conference 2020
DOI: 10.1109/globecom42002.2020.9322237
|View full text |Cite
|
Sign up to set email alerts
|

Continuous Multi-objective Zero-touch Network Slicing via Twin Delayed DDPG and OpenAI Gym

Abstract: Artificial intelligence (AI)-driven zero-touch network slicing (NS) is a new paradigm enabling the automation of resource management and orchestration (MANO) in multi-tenant beyond 5G (B5G) networks. In this paper, we tackle the problem of cloud-RAN (C-RAN) joint slice admission control and resource allocation by first formulating it as a Markov decision process (MDP). We then invoke an advanced continuous deep reinforcement learning (DRL) method called twin delayed deep deterministic policy gradient (TD3) to … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
23
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

3
6

Authors

Journals

citations
Cited by 32 publications
(23 citation statements)
references
References 11 publications
0
23
0
Order By: Relevance
“…The DNNs structure for the actor-critic networks and target networks are the same. We have set the hyperparameters following extensive experiments [27]. The evaluation computes every 20000 iterations concerning the average return over the best 3 of 5 episodes.…”
Section: Numerical Resultsmentioning
confidence: 99%
“…The DNNs structure for the actor-critic networks and target networks are the same. We have set the hyperparameters following extensive experiments [27]. The evaluation computes every 20000 iterations concerning the average return over the best 3 of 5 episodes.…”
Section: Numerical Resultsmentioning
confidence: 99%
“…Table I presents the network parameters. We set hyperparameters of DNNs through extensive experiments [25]- [26] and adopt a similar architecture for both Actor-Critic and target DNN models. We use 5 hidden layers and 128 units per layer with batch size 128.…”
Section: Numerical Resultsmentioning
confidence: 99%
“…There exist 10 APs and a maximum of 17 registered subscribers that are assigned randomly to different slices in each decision time Table I presents the network parameters. We set hyperparameters of DNNs through extensive experiments [25] and adopt a similar architecture for both Actor-Critic and target DNN models. We use 5 hidden layers and 128 units per layer with batch size 128.…”
Section: Numerical Resultsmentioning
confidence: 99%