2021
DOI: 10.1088/1742-6596/1885/4/042070
|View full text |Cite
|
Sign up to set email alerts
|

Quadric Lyapunov Algorithm for Stochastic Networks Optimization with Q-learning Perspective

Abstract: In this article, we investigate stochastic networks optimization using Quadric Lyapunov Algorithm (QLA) with Q-learning perspective. We proposed firstly a model of stochastic queueing networks with power constraints. QLA is then proposed aiming at minimizing an expression containing Lyapunov drift. Based on the analysed similarity between QLA and Q-learning, we show the possibility and feasibility of Q-learning. Simulation of a simple queue network model is carried out, and results using both QLA and Q-learnin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 4 publications
0
2
0
Order By: Relevance
“…After decades of development, RL technology has many achievements, such as Q-learning, dynamic programming, Policy Gradients, Deep-Q-Network, etc. [13][14][15][16][17][18][19][20][21][22][23][24]. In essence, RL is a process in which the agent learns itself in an unknown environment under defined rules.…”
Section: Introductionmentioning
confidence: 99%
“…After decades of development, RL technology has many achievements, such as Q-learning, dynamic programming, Policy Gradients, Deep-Q-Network, etc. [13][14][15][16][17][18][19][20][21][22][23][24]. In essence, RL is a process in which the agent learns itself in an unknown environment under defined rules.…”
Section: Introductionmentioning
confidence: 99%
“…After decades of development, RL has achieved many achievements, such as Q-learning, Dynamic Programing, Policy Gradients, Deep-Q-Network, etc. [9][10][11][12][13][14][15][16].…”
Section: Introductionmentioning
confidence: 99%