2021
DOI: 10.3390/s21248306
|View full text |Cite
|
Sign up to set email alerts
|

Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning

Abstract: As the dynamicity of the traffic increases, the need for self-network operation becomes more evident. One of the solutions that might bring cost savings to network operators is the dynamic capacity management of large packet flows, especially in the context of packet over optical networks. Machine Learning, particularly Reinforcement Learning, seems to be an enabler for autonomicity as a result of its inherent capacity to learn from experience. However, precisely because of that, RL methods might not be able t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

4
1

Authors

Journals

citations
Cited by 9 publications
(11 citation statements)
references
References 24 publications
0
11
0
Order By: Relevance
“…A key novel contribution of this paper is the design and evaluation of the DRL-based flow routing agent, responsible of managing the routing of a target flow. From the lessons learnt in [6], a number of modules and workflows are presented to allow efficient and robust DRL-based routing immediately after the flow is provisioned and operation starts. Note that the DRL engine is able to learn and improve actions as soon as it gets experience from operation.…”
Section: Related Work and Contributionsmentioning
confidence: 99%
See 4 more Smart Citations
“…A key novel contribution of this paper is the design and evaluation of the DRL-based flow routing agent, responsible of managing the routing of a target flow. From the lessons learnt in [6], a number of modules and workflows are presented to allow efficient and robust DRL-based routing immediately after the flow is provisioned and operation starts. Note that the DRL engine is able to learn and improve actions as soon as it gets experience from operation.…”
Section: Related Work and Contributionsmentioning
confidence: 99%
“…The DRL agent consists of two inter-related blocks: i) the learning agent is in charge of learning the best actions to be taken based on the current state and the received reward; and ii) the environment that is in charge of computing the state based on the observed traffic and current combination of routes (proportions), as well as the obtained reward based on the measured e2e delay for the selected routes. Without loss of generality, we assume that the DRL agent operates under the life-cycle presented in [6], where models are pre-trained offline with generic data and refined during online learning, thus achieving both decision-making robustness and high performance from the beginning of flow operation. In Fig.…”
Section: Distributed Network Intelligence For Autonomous Flow Routingmentioning
confidence: 99%
See 3 more Smart Citations