2019
DOI: 10.1002/acs.2981
|View full text |Cite
|
Sign up to set email alerts
|

Experience replay–based output feedback Q‐learning scheme for optimal output tracking control of discrete‐time linear systems

Abstract: This paper focuses on solving the adaptive optimal tracking control problem for discrete-time linear systems with unknown system dynamics using output feedback. A Q-learning-based optimal adaptive control scheme is presented to learn the feedback and feedforward control parameters of the optimal tracking control law. The optimal feedback parameters are learned using the proposed output feedback Q-learning Bellman equation, whereas the estimation of the optimal feedforward control parameters is achieved using a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
4
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 56 publications
0
4
0
Order By: Relevance
“…In this paper, we propose a tracking control approach that utilizes a static output feedback multistep Q-learning algorithm in conjunction with state reconstruction techniques. A separate adaptation mechanism was introduced in [38] to estimate unknown feedforward tracking terms. However, the static OPFB design we propose owes its popularity to its simplicity in terms of structure.…”
Section: Introductionmentioning
confidence: 99%
“…In this paper, we propose a tracking control approach that utilizes a static output feedback multistep Q-learning algorithm in conjunction with state reconstruction techniques. A separate adaptation mechanism was introduced in [38] to estimate unknown feedforward tracking terms. However, the static OPFB design we propose owes its popularity to its simplicity in terms of structure.…”
Section: Introductionmentioning
confidence: 99%
“…Recently, reinforcement learning has emerged as an effective method to solve optimal control problems without using system model information 8‐10 . Q‐learning is a type of reinforcement learning algorithm, which is completely online in nature and does not use any prior information of system dynamics.…”
Section: Introductionmentioning
confidence: 99%
“…In recent decades, a large number of researches on the problem of H ∞ control have emerged . Meanwhile, the tracking control problem has been widely studied as it appears in many practical control applications . Combining tracking control with H ∞ control, it can attenuate the effects of disturbances on system and ensure the system trajectory tracks the reference signal.…”
Section: Introductionmentioning
confidence: 99%
“…[3][4][5][6] Meanwhile, the tracking control problem has been widely studied as it appears in many practical control applications. [7][8][9] Combining tracking control with H ∞ control, it can attenuate the effects of disturbances on system and ensure the system trajectory tracks the reference signal. Therefore, the research on H ∞ tracking control is meaningful.…”
Section: Introductionmentioning
confidence: 99%