2017
DOI: 10.1137/16m1086261
|View full text |Cite
|
Sign up to set email alerts
|

Continuous-Time Markov Decision Processes with Exponential Utility

Abstract: In this paper, we consider a continuous-time Markov decision process (CTMDP) in Borel spaces, where the certainty equivalent with respect to the exponential utility of the total undiscounted cost is to be minimized. The cost rate is nonnegative. We establish the optimality equation. Under the compactness-continuity condition, we show the existence of a deterministic stationary optimal policy. We reduce the risk-sensitive CTMDP problem to an equivalent risk-sensitive discrete-time Markov decision process, which… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
21
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
7

Relationship

2
5

Authors

Journals

citations
Cited by 32 publications
(22 citation statements)
references
References 32 publications
(121 reference statements)
1
21
0
Order By: Relevance
“…For ease of reference, we present the relevant notations and facts about the risk-sensitive problem for a DTMDP. The proofs of the presented statements can be found in [16] or [27]. Standard description of a DTMDP can be found in e.g., [17,22].…”
Section: A Appendixmentioning
confidence: 99%
“…For ease of reference, we present the relevant notations and facts about the risk-sensitive problem for a DTMDP. The proofs of the presented statements can be found in [16] or [27]. Standard description of a DTMDP can be found in e.g., [17,22].…”
Section: A Appendixmentioning
confidence: 99%
“…An early work on this topic seems to be [18], which obtained verification theorems and solved in closed-form meaningful examples of problems over a fixed time duration. In the recent years, there have been reviving interests in risk-sensitive CTMDPs, see [5,6,21] for problems with a finite horizon, [23] for problems over an infinite horizon, [5,16,22] for problems with average criteria, and [2] for an optimal stopping problem with a more general utility function than the exponential one.…”
Section: Introductionmentioning
confidence: 99%
“…We denote by the set of all randomized history-dependent policies, by r m the set of all randomized Markov policies, by d m the set of all (deterministic) Markov policies, and by F the set of all stationary policies. Obviously, ⊃ r m ⊃ d m ⊃ F. For any initial distribution γ on S and policy π ∈ , we introduce the unique probability measure P π γ on ( , F) as in [12,13,28]. Let E π γ be its corresponding expectation operator.…”
Section: The Risk-sensitive Average Optimality Problemmentioning
confidence: 99%
“…In comparison, there has been limited work on risk-sensitive continuous-time Markov decision processes (CTMDPs). For background on risk-sensitive CTMDPs, see [7,11,22] for the risk-sensitive (infinite-horizon) discounted criterion, [7,22,23,25,26,27] for the risk-sensitive (long-run) average criterion, [7,9,14,17,18,23,24] for the risk-sensitive finite-horizon criterion, and [8,28] for the risk-sensitive total cost criterion. In this paper we will further consider the risksensitive average criterion for CTMDPs studied in [7,22,23,25,26,27].…”
Section: Introductionmentioning
confidence: 99%