2015
DOI: 10.1109/tcyb.2015.2417170
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement-Learning-Based Robust Controller Design for Continuous-Time Uncertain Nonlinear Systems Subject to Input Constraints

Abstract: The design of stabilizing controller for uncertain nonlinear systems with control constraints is a challenging problem. The constrained-input coupled with the inability to identify accurately the uncertainties motivates the design of stabilizing controller based on reinforcement-learning (RL) methods. In this paper, a novel RL-based robust adaptive control algorithm is developed for a class of continuous-time uncertain nonlinear systems subject to input constraints. The robust control problem is converted to t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

3
136
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
9

Relationship

1
8

Authors

Journals

citations
Cited by 342 publications
(139 citation statements)
references
References 46 publications
3
136
0
Order By: Relevance
“…A policy iteration algorithm was introduced for infinite horizon optimal control of nonlinear systems using ADP in [67]. In [68], a reinforcement learning method was introduced for the stabilizing control of uncertain nonlinear…”
Section: Nn Based Adaptive Dynamicmentioning
confidence: 99%
“…A policy iteration algorithm was introduced for infinite horizon optimal control of nonlinear systems using ADP in [67]. In [68], a reinforcement learning method was introduced for the stabilizing control of uncertain nonlinear…”
Section: Nn Based Adaptive Dynamicmentioning
confidence: 99%
“…It brings additional challenge for the control design. Some researches have been done for the input constraint problem [9][10][11]. In [12,13], an auxiliary system is novelly introduced to handle the constraint effect in a class of uncertain multiple-input multiple-output nonlinear system.…”
Section: Introductionmentioning
confidence: 99%
“…The general structure for implementing the ADP algorithm is actor-critic architecture, where the actor performs actions by interacting with its environment, and the critic evaluates actions and offers feedback information to the actor, leading to the improvement in performance of the subsequent actor [4]. Since then, various ADP methods have been developed [4][5][6][7][8][9][10][11][12][13][14]. However, most of the existing ADP-related results require an exact knowledge of nonlinear dynamics, which is often unavailable, thus these ADP algorithms are intractable to real-time control applications.…”
Section: Introductionmentioning
confidence: 99%