2014
DOI: 10.1002/rnc.3152
|View full text |Cite
|
Sign up to set email alerts
|

Approximate dynamic programming via iterated Bellman inequalities

Abstract: SUMMARYIn this paper we introduce new methods for finding functions that lower bound the value function of a stochastic control problem, using an iterated form of the Bellman inequality. Our method is based on solving linear or semidefinite programs, and produces both a bound on the optimal objective, as well as a suboptimal policy that appears to work very well. These results extend and improve bounds obtained in a previous paper using a single Bellman inequality condition. We describe the methods in a genera… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
116
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 70 publications
(116 citation statements)
references
References 66 publications
(88 reference statements)
0
116
0
Order By: Relevance
“…This value-function-based method is called DP, and a variety of topics on stochastic optimal control and DP are well-addressed by [5][6][7][8]. A large class of stochastic optimal control problems deal with the dynamics of the form in Equation (1) and are concerned with finding a state-feedback control policy:…”
Section: Preliminariesmentioning
confidence: 99%
See 4 more Smart Citations
“…This value-function-based method is called DP, and a variety of topics on stochastic optimal control and DP are well-addressed by [5][6][7][8]. A large class of stochastic optimal control problems deal with the dynamics of the form in Equation (1) and are concerned with finding a state-feedback control policy:…”
Section: Preliminariesmentioning
confidence: 99%
“…where T is the Bellman operator (see e.g., [5]), whose domain and codomain are both function spaces mapping X onto…”
Section: Preliminariesmentioning
confidence: 99%
See 3 more Smart Citations