2021
DOI: 10.48550/arxiv.2109.13132
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Optimization Landscape of Gradient Descent for Discrete-time Static Output Feedback

Abstract: In this paper, we analyze the optimization landscape of the gradient descent method for static output feedback (SOF) control of discrete-time linear time-invariant systems with quadratic cost. The SOF setting can be quite common, for example, when there are unmodeled hidden states in the underlying process. We first identify several important properties of the SOF cost function, including coercivity, L-smoothness, and M -Lipschitz continuous Hessian. Based on these results, we show that when the observation ma… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 22 publications
0
6
0
Order By: Relevance
“…Recent theoretical results on PO for particular classes of control synthesis problems, some of which are discussed in this survey, not only are exciting but also lead to a new research thrust at the interface of control theory and machine learning. This survey includes control synthesis related to linear quadratic regulator (LQR) theory (35)(36)(37)(38)(39)(40)(41)(42)(43)(44), stabilization (45)(46)(47), linear robust/risk-sensitive control (48)(49)(50)(51)(52)(53)(54)(55), Markov jump linear quadratic control (56-59), Lur'e system control (60), output feedback control (61)(62)(63)(64)(65)(66)(67), and dynamic filtering (68). Surprisingly, some of these strong global convergence results for PO have been obtained in the absence of convexity in the design objective and/or the underlying feasible set.…”
Section: Introductionmentioning
confidence: 99%
“…Recent theoretical results on PO for particular classes of control synthesis problems, some of which are discussed in this survey, not only are exciting but also lead to a new research thrust at the interface of control theory and machine learning. This survey includes control synthesis related to linear quadratic regulator (LQR) theory (35)(36)(37)(38)(39)(40)(41)(42)(43)(44), stabilization (45)(46)(47), linear robust/risk-sensitive control (48)(49)(50)(51)(52)(53)(54)(55), Markov jump linear quadratic control (56-59), Lur'e system control (60), output feedback control (61)(62)(63)(64)(65)(66)(67), and dynamic filtering (68). Surprisingly, some of these strong global convergence results for PO have been obtained in the absence of convexity in the design objective and/or the underlying feasible set.…”
Section: Introductionmentioning
confidence: 99%
“…This is also known as Partially Observable Markov Decision Process (POMDP) in the Markovian system setting [18]. Some recent works have studied static output-feedback (SOF) controllers to optimize a linear quadratic cost function [10], [19]- [21]. Unlike state-feedback LQR problems, it is shown that policy gradient methods are unlikely to find the globally optimal SOF controller.…”
Section: Introductionmentioning
confidence: 99%
“…This paper takes a step further to analyze the optimization landscape of Dynamic output-feedback LQR (dLQR). Unlike the vanilla LQR and the SOF that use static feedback policies [10], [19]- [21], the problem of dLQR searches over the set of dynamic controllers, which has rich yet complicated landscape properties. The recent work [25], [26] has analyzed the structure of optimal dynamic controllers for the classical Linear Quadratic Gaussian (LQG) control problem.…”
Section: Introductionmentioning
confidence: 99%
“…To this end, we analyze the classic single-sample AC for solving the Linear Quadratic Regulator (LQR), a fundamental control task which is commonly employed as a testbed to explore the behavior and limits of RL algorithms under continuous state-action spaces (Fazel et al 2018;Yang et al 2019;Tu and Recht 2018;Krauth, Tu, and Recht 2019;Duan, Li, and Zhao 2021). In the LQR case, the Q-function is a linear function of the quadratic form of state and action.…”
Section: Introductionmentioning
confidence: 99%