Proceedings of the 2005, American Control Conference, 2005.
DOI: 10.1109/acc.2005.1470149
|View full text |Cite
|
Sign up to set email alerts
|

Stagewise Newton, differential dynamic programming, and neighboring optimum control for neural-network learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
9
0

Publication Types

Select...
4
1

Relationship

2
3

Authors

Journals

citations
Cited by 11 publications
(9 citation statements)
references
References 21 publications
0
9
0
Order By: Relevance
“…Stage-wise Newton (SN) method [26] and differential dynamic programming (DDP) [27], [28] are the two standard methods for this purpose. The pure SN method [26] sequentially optimizes a second-order model of the Hamiltonian H k px k , u k , λ k`1 q " L k px k , u k q`λ T k`1 f k px k , u k q while DDP sequentially optimizes a local model of the Hamilton-Jacobi-Bellman value function V k px k , u k q defined recursively by Vk px k q " min…”
Section: Stage-wise Newton and Differential Dynamic Programmingmentioning
confidence: 99%
“…Stage-wise Newton (SN) method [26] and differential dynamic programming (DDP) [27], [28] are the two standard methods for this purpose. The pure SN method [26] sequentially optimizes a second-order model of the Hamiltonian H k px k , u k , λ k`1 q " L k px k , u k q`λ T k`1 f k px k , u k q while DDP sequentially optimizes a local model of the Hamilton-Jacobi-Bellman value function V k px k , u k q defined recursively by Vk px k q " min…”
Section: Stage-wise Newton and Differential Dynamic Programmingmentioning
confidence: 99%
“…On the other hand, such nice sparsity may disappear when weight-sharing and weight-pruning are applied (as usual in optimal control [8]) so that all the terminal parameters 0N-IN are shared among the terminal states yN. In this way, MLPlearning exhibits a great deal of structure.…”
Section: Discussionmentioning
confidence: 96%
“…denotes a differentiable state-transition, function of nonlinear dynamics, and x1, the "before-node" net input to node j at layer s, depends on only a subset of all decisions taken at stage s-1. In spite of this distinction and others, using a vector of states as a basic ingredient allows us to adopt analogous formulas available in the optimal control theory (see [8]). The key concept behind the theory resides in stagewise implementation; in fact, first-order BP is essentially a simplified stagewise optimalcontrol gradient formula developed in early 1960s [6].…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…Learning with multilayer-perceptron (MLP) neural networks is a multiple N -stage decision-making (or discretetime optimal control) problem [1]. The optimal-control theory available in control engineering dictates to us a variety of elaborate learning schemes in the context of N -layered MLP-learning with P s nodes at layer s (s = 1, ..., N ).…”
Section: Introductionmentioning
confidence: 99%