1976
DOI: 10.1007/bf01441962
|View full text |Cite
|
Sign up to set email alerts
|

Approximations and computational methods for Optimal Stopping and Stochastic Impulsive Control problems

Abstract: The paper treats a computational method for the Optimal Stopping and Stochastic Impulsive Control problem for a diffusion. In the latter problem control acts only intermittently since there is a basic positive "transaction" cost to be paid at each instant that the control acts. For each h > 0, a controlled Markov chain is constructed, whose continuous time interpolations are a natural approximation to the diffusion, for both the optimal stopping and impulsive control situations. The solutions to the optimal st… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

1981
1981
2014
2014

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(1 citation statement)
references
References 8 publications
0
1
0
Order By: Relevance
“…The algorithm converges to the optimal value function at the bottom of Figure 4 and the optimal policy is described by d D U u = 0 0334 0.0614, 0.0853, 0.1343 . Although one can discretize the problem using the finite difference schemes proposed in Kushner (1976) and then brute force the solution using value or policy iteration, this methodology is not expected to perform well on impulse control problems because of the discontinuities in the state evolution that translate to insufficient smoothness in the value function for accelerated convergence. However, for illustration, comparing our method with that of using policy iteration on the discretized Markov chain, our method converges after 22 iterations in 0.176 seconds on a 10,000 point grid, whereas policy iteration in the controlled Markov chain case converges after 703 iterations in 1,328 seconds on the same grid.…”
Section: Analysis Of Resultsmentioning
confidence: 98%
“…The algorithm converges to the optimal value function at the bottom of Figure 4 and the optimal policy is described by d D U u = 0 0334 0.0614, 0.0853, 0.1343 . Although one can discretize the problem using the finite difference schemes proposed in Kushner (1976) and then brute force the solution using value or policy iteration, this methodology is not expected to perform well on impulse control problems because of the discontinuities in the state evolution that translate to insufficient smoothness in the value function for accelerated convergence. However, for illustration, comparing our method with that of using policy iteration on the discretized Markov chain, our method converges after 22 iterations in 0.176 seconds on a 10,000 point grid, whereas policy iteration in the controlled Markov chain case converges after 703 iterations in 1,328 seconds on the same grid.…”
Section: Analysis Of Resultsmentioning
confidence: 98%