2017
DOI: 10.1007/978-981-10-4080-1_5
|View full text |Cite
|
Sign up to set email alerts
|

Nonlinear Neuro-Optimal Tracking Control via Stable Iterative Q-Learning Algorithm

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
3
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
2
2

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 43 publications
0
3
0
Order By: Relevance
“…To address this problem, approximation-based control strategies have been developed based on fuzzy logic system [19] and neural networks (NNs). NNs, with their expressive power of approximation widely recognized, have been utilized to represent direct dynamics, inverse dynamics, or a certain nonlinear mapping [20]- [22]. In general, NNsbased dynamics compensation can be classified into two main categories: either NN is used to produce control inputs for a feedback controller that is designed based on a nominal robot dynamic model, or NN directly estimates the dynamic system inversion.…”
Section: Introductionmentioning
confidence: 99%
“…To address this problem, approximation-based control strategies have been developed based on fuzzy logic system [19] and neural networks (NNs). NNs, with their expressive power of approximation widely recognized, have been utilized to represent direct dynamics, inverse dynamics, or a certain nonlinear mapping [20]- [22]. In general, NNsbased dynamics compensation can be classified into two main categories: either NN is used to produce control inputs for a feedback controller that is designed based on a nominal robot dynamic model, or NN directly estimates the dynamic system inversion.…”
Section: Introductionmentioning
confidence: 99%
“…For many traditional iterative ADP algorithms, it is required to build the model of nonlinear systems and then perform the ADP algorithms to derive an improved control policy [43][44][45][46][47][48][49][50]. These iterative ADP algorithms are denoted as "model-based ADP algorithms".…”
Section: Introductionmentioning
confidence: 99%
“…In [11], interval type-2 fuzzy logic control and actor-critic RL algorithms with one-order digital low-pass filters are used to overcome the difficulties-caused by the nonlinearities and uncertainties of the system and the working environment-of accurate trajectory tracking control. An iterative adaptive dynamic programming (DP) technique is adopted to build the stable iterative Q-learning algorithm for achieving nonlinear neuro-optimal tracking control in [12]. To control a robotic manipulator with unknown parameters and dead zones, a neural network model for RL is described in [13].…”
mentioning
confidence: 99%