2019
DOI: 10.1007/s11432-018-9865-9
|View full text |Cite
|
Sign up to set email alerts
|

Online adaptive Q-learning method for fully cooperative linear quadratic dynamic games

Abstract: A model-based offline policy iteration (PI) algorithm and a model-free online Q-learning algorithm are proposed for solving fully cooperative linear quadratic dynamic games. The PI-based adaptive Q-learning method can learn the feedback Nash equilibrium online using the state samples generated by behavior policies, without sending inquiries to the system model. Unlike the existing Q-learning methods, this novel Q-learning algorithm executes both policy evaluation and policy improvement in an adaptive manner. W… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 12 publications
(1 citation statement)
references
References 32 publications
(56 reference statements)
0
1
0
Order By: Relevance
“…Zhang et al [42] exposited the data-driven ADP al-gorithm to study the cooperative game problem with constrained input. Li et al [43] expounded a new adaptive Q-learning algorithm to solve cooperative linear quadratic dynamic game. Mu et al [1] transformed the cooperative differential game problem into the optimal control problem by defining the global performance index function and successfully applied cooperative game to power system.…”
Section: Introductionmentioning
confidence: 99%
“…Zhang et al [42] exposited the data-driven ADP al-gorithm to study the cooperative game problem with constrained input. Li et al [43] expounded a new adaptive Q-learning algorithm to solve cooperative linear quadratic dynamic game. Mu et al [1] transformed the cooperative differential game problem into the optimal control problem by defining the global performance index function and successfully applied cooperative game to power system.…”
Section: Introductionmentioning
confidence: 99%