2009
DOI: 10.1016/j.neunet.2009.03.008
|View full text |Cite
|
Sign up to set email alerts
|

Neural network approach to continuous-time direct adaptive optimal control for partially unknown nonlinear systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
352
0

Year Published

2012
2012
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 562 publications
(352 citation statements)
references
References 20 publications
0
352
0
Order By: Relevance
“…[37]. The IRL-PI algorithm proposed converges uniformly to the optimal control solution on trajectories originating in Ω, i.e.…”
Section: (Policy Evaluation Step) Solve For the Value Functionmentioning
confidence: 97%
See 1 more Smart Citation
“…[37]. The IRL-PI algorithm proposed converges uniformly to the optimal control solution on trajectories originating in Ω, i.e.…”
Section: (Policy Evaluation Step) Solve For the Value Functionmentioning
confidence: 97%
“…Consider the following affine in the control input nonlinear system with quadratic cost function applied in [37], which has the dynamics…”
Section: Nonlinear System Examplementioning
confidence: 99%
“…. , N. Remark 3: According to (42), it is obvious to observe that the approximate control policies of the N isolated subsystems can be derived directly based on the trained critic networks. Therefore, unlike the traditional actor-critic architecture, the action neural networks are not required any more.…”
Section: B Implementation Procedures Via Neural Networkmentioning
confidence: 99%
“…Abu-Khalaf and Lewis [41] derived an offline optimal control scheme for nonlinear systems with saturating actuators. Then, Vrabie and Lewis [42] and Vamvoudakis and Lewis [43] used online policy iteration algorithm to study the infinite horizon optimal control of continuous-time nonlinear systems, respectively. The former was performed based on the sequential updates of two neural networks, namely, critic network and action network, whereas in the latter, the two networks were trained simultaneously.…”
Section: Introductionmentioning
confidence: 99%
“…In particular, biological experiments show that the dopamine neurotransmitter acts as a reinforcement signal which favors learning at the neuron level [12]. Based on reinforcement learning, a control signal can be generated for an agent to interact with unknown environments.…”
Section: Introductionmentioning
confidence: 99%