2022
DOI: 10.48550/arxiv.2202.05404
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Regularized Q-learning

Abstract: Q-learning is widely used algorithm in reinforcement learning community. Under the lookup table setting, its convergence is well established. However, its behavior is known to be unstable with the linear function approximation case. This paper develops a new Q-learning algorithm that converges when linear function approximation is used. We prove that simply adding an appropriate regularization term ensures convergence of the algorithm. We prove its stability using a recent analysis tool based on switching syst… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 23 publications
(47 reference statements)
0
1
0
Order By: Relevance
“…The analysis in [22] includes both asymptotic and finitetime natures, while some parameters in the bounds are not explicit. The paper [21] studies asymptotic convergence of Q-learning [4] through a continuous-time switched linear system model [27], [28].…”
Section: B Related Work 1) Finite-time Analysis Of Td-learningmentioning
confidence: 99%
“…The analysis in [22] includes both asymptotic and finitetime natures, while some parameters in the bounds are not explicit. The paper [21] studies asymptotic convergence of Q-learning [4] through a continuous-time switched linear system model [27], [28].…”
Section: B Related Work 1) Finite-time Analysis Of Td-learningmentioning
confidence: 99%