2012
DOI: 10.1002/rnc.2894
|View full text |Cite
|
Sign up to set email alerts
|

Quadratic approximate dynamic programming for input‐affine systems

Abstract: SUMMARYWe consider the use of quadratic approximate value functions for stochastic control problems with input‐affine dynamics and convex stage cost and constraints. Evaluating the approximate dynamic programming policy in such cases requires the solution of an explicit convex optimization problem, such as a quadratic program, which can be carried out efficiently. We describe a simple and general method for approximate value iteration that also relies on our ability to solve convex optimization problems, in th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
20
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 26 publications
(20 citation statements)
references
References 39 publications
0
20
0
Order By: Relevance
“…Another tool used in portfolio optimization applications of this paper is a special class of approximate value function methods [27][28][29][30]. In general, stochastic optimal control problems can be solved by utilizing state value functions, which estimate performance at a given state.…”
Section: Initialize Parameter θ Of the Search Distribution π(·|θ)mentioning
confidence: 99%
See 2 more Smart Citations
“…Another tool used in portfolio optimization applications of this paper is a special class of approximate value function methods [27][28][29][30]. In general, stochastic optimal control problems can be solved by utilizing state value functions, which estimate performance at a given state.…”
Section: Initialize Parameter θ Of the Search Distribution π(·|θ)mentioning
confidence: 99%
“…in state-space format, it is necessary to define the state and control input together with the performance index that is used as an optimization criterion. To do this, we follow the research of Boyd et al [1,28,30]. We define the state vector as the collection of the portfolio positions.…”
Section: Machine Learning and Control Based Portfolio Optimizationmentioning
confidence: 99%
See 1 more Smart Citation
“…For the special case when the dynamics of the system are linear, Dynamic Programming (DP) gives a complete and explicit solution to the problem, because the one-step state cost and the value/cost function in this case are quadratic. 16 For the general nonlinear control problem, DP is difficult to carry out and ADP designs are not systematic.…”
mentioning
confidence: 99%
“…This is called a model-free approach, because it does not need any a priori model information at the beginning of the algorithm nor on-line identification of nonlinear systems, but only the on-line identified linear model. This control approach was inspired by the ideas and solutions given by several articles [16][17][18][19][20] . It starts with the selection of the value/cost function in a systematic way, 16 and follows by the Linear Approximate Dynamic Programming (LADP) model-free adaptive control approach.…”
mentioning
confidence: 99%