2023
DOI: 10.1109/tac.2022.3145632
|View full text |Cite
|
Sign up to set email alerts
|

Linear Quadratic Control Using Model-Free Reinforcement Learning

Abstract: In this paper, we consider Linear Quadratic (LQ) control problem with process and measurement noises. We analyze the LQ problem in terms of the average cost and the structure of the value function. We assume that the dynamics of the linear system is unknown and only noisy measurements of the state variable are available. Using noisy measurements of the state variable, we propose two model-free iterative algorithms to solve the LQ problem. The proposed algorithms are variants of policy iteration routine where t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
16
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(16 citation statements)
references
References 35 publications
0
16
0
Order By: Relevance
“…In order to better explain Algorithm 2, we can divide it into the following three steps: First of all, prove that Ĝi can be uniquely determined in each iteration. Lemma 5 explains that the solutions obtained from the Bellman Equation ( 26) based on the average off-policy Q-learning method are equivalent to those obtained from the model-based Bellman Equation (18). Equation (46) of Algorithm 2 is derived from the transformation of Equation ( 26), and we show that the procedure in Algorithm 2 guarantees the existence of the solution Ĝi .…”
Section: Lemma 4 ([30]mentioning
confidence: 86%
See 2 more Smart Citations
“…In order to better explain Algorithm 2, we can divide it into the following three steps: First of all, prove that Ĝi can be uniquely determined in each iteration. Lemma 5 explains that the solutions obtained from the Bellman Equation ( 26) based on the average off-policy Q-learning method are equivalent to those obtained from the model-based Bellman Equation (18). Equation (46) of Algorithm 2 is derived from the transformation of Equation ( 26), and we show that the procedure in Algorithm 2 guarantees the existence of the solution Ĝi .…”
Section: Lemma 4 ([30]mentioning
confidence: 86%
“…By utilizing an average cost function, Reference [17] tackled the output regulation problem for linear systems with unknown dynamics. Data-driven average Q-learning algorithms can also handle linear quadratic control problems, especially when there are un-measurable stochastic disturbances [18,19].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…An iterative Linear Quadratic Regulator (iLQR) optimal control technique based on the dynamic modeling of the quadrotors was developed to achieve the leader-follower formation (Jasim and Gu 2019). Among the array of control methodologies discussed earlier, Reinforcement Learning (RL)-based control approaches offer an end-to-end solution for motion control (Yaghmaie et al 2023) and obstacle avoidance (Sadhukhan and Selmic 2021) during formation.…”
Section: Introductionmentioning
confidence: 99%
“…However, adding a terminal constraints set increases computational demand at each step. Thus, we proposed a LQG-based approach (Yaghmaie et al, 2022) to reduce the disturbance and increase system stability.…”
Section: Introductionmentioning
confidence: 99%