2018
DOI: 10.1007/s12555-016-0711-5
|View full text |Cite
|
Sign up to set email alerts
|

Model-free Adaptive Dynamic Programming Based Near-optimal Decentralized Tracking Control of Reconfigurable Manipulators

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 51 publications
(11 citation statements)
references
References 56 publications
0
11
0
Order By: Relevance
“…Using (11) and (12), the approximated optimal control strategy and positive definition function are obtained…”
Section: The Aperiodic Critic Nns Designmentioning
confidence: 99%
See 1 more Smart Citation
“…Using (11) and (12), the approximated optimal control strategy and positive definition function are obtained…”
Section: The Aperiodic Critic Nns Designmentioning
confidence: 99%
“…The ADP technique has been applied successfully to strict-feedback non-linear systems [5,6], discrete-time non-linear systems [7,8] and nonlinear switched systems [9,10]. Very recently, the ADP technique was extended to derive the decentralised tracking controller for non-linear interconnected systems [11,12]. [11] used an observer-critic structure-based to reconstruct unknown system dynamics and solve the coupled Hamilton-Jacobi-Bellman (HJB) equation, respectively.…”
Section: Introductionmentioning
confidence: 99%
“…According to [38,45], the HJBE (8) can be solved by time-triggered ADP algorithm. However, as mentioned in [46], the time-triggered optimal control strategies do not only suffer from heavy computational burden and communication, but also waste limited energy resource.…”
Section: Dynamic Model and Preliminariesmentioning
confidence: 99%
“…2. Unlike existing time-triggered control methods [37,38] which ignored the conservation of limited energy resource, in this paper, a novel compensator-critic structure-based event-triggered decentralized tracking control method for MRMs is proposed. It does not only make the actual trajectory of each joint module follow its desired one, but also reduce the computational burden, save the communication, and energy consumption simultaneously.…”
Section: Introductionmentioning
confidence: 99%
“…Reinforcement learning (RL) is a popular algorithm for determining a policy that optimizes a user‐defined cost function through interactions with an environment 1‐7 . There are several applications of RL in chemical process engineering, such as a model‐free learning controller for pH control, 8 a policy gradient approach for batch bioprocess systems, 9 and approximate dynamic programming (ADP) for control of fed‐batch bioreactors, 10 control of proppant concentrations inside a fracture for hydraulic fracturing, 11 and control of alkali‐surfactant‐polymer flooding for oil recovery 12 .…”
Section: Introductionmentioning
confidence: 99%