2020
DOI: 10.48550/arxiv.2004.14288
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Actor-Critic Reinforcement Learning for Control with Stability Guarantee

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
10
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

5
2

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 26 publications
0
10
0
Order By: Relevance
“…Motivated by the works in [20], [25], [26], [27], we propose to incorporate the theoretical result in Theorem 1 to formulate a constrained optimisation problem, based on SAC [17]. First of all, a Lyapunov candidate needs to be selected at the first instance.…”
Section: Lyapunov-based Reinforcement Learningmentioning
confidence: 99%
See 2 more Smart Citations
“…Motivated by the works in [20], [25], [26], [27], we propose to incorporate the theoretical result in Theorem 1 to formulate a constrained optimisation problem, based on SAC [17]. First of all, a Lyapunov candidate needs to be selected at the first instance.…”
Section: Lyapunov-based Reinforcement Learningmentioning
confidence: 99%
“…[19] proposes a straightforward approach to construct the Lyapunov functions for nonlinear systems using DNNs. Recently, the asymptotic stability in model-free RL is given for robotic control tasks in [20]. Inspired by the works [19], [20], we will also parametrise the Lyapunov function as a DNN and learn the parameters from samples.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Analysis based on the Lyapunov theorem needs to construct a Lyapunov function to ensure the stability [10]. The Lyapunov stability methods have been widely used in the field of control engineering [10] and recently well investigated for RL algorithms to ensure the stability [11]. As another method, the input-output stability is also used to achieve the same purpose [12].…”
Section: Introductionmentioning
confidence: 99%
“…However, the classical control methods rely on the full or partial knowledge of the system dynamics to design controllers and are largely limited to systems with simple dynamics. Thus, it is a natural move to combine RL with control theory to develop learning control methods with a stability guarantee [5,6]. This paper was not presented at any IFAC meeting.…”
Section: Introductionmentioning
confidence: 99%