2004
DOI: 10.1081/sta-120037268
|View full text |Cite
|
Sign up to set email alerts
|

Markov Decision Processes with Asymptotic Average Failure Rate Constraint

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2013
2013
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 14 publications
(6 citation statements)
references
References 7 publications
0
6
0
Order By: Relevance
“…We also present examples of Dynkin's formulae and boundary value problems for controlled additive functionals (CAFs) and controlled geometric Markov renewal chains (CGMRCs). In the literature a CAF is usually called the Markov decision process (see, for example, [3], [7], [8], and [9]). We also mention that the discrete-time case of the stochastic optimal control has been considered in [6].…”
Section: Optimal Control and The Hamilton-jacobi-bellman Equation Formentioning
confidence: 99%
“…We also present examples of Dynkin's formulae and boundary value problems for controlled additive functionals (CAFs) and controlled geometric Markov renewal chains (CGMRCs). In the literature a CAF is usually called the Markov decision process (see, for example, [3], [7], [8], and [9]). We also mention that the discrete-time case of the stochastic optimal control has been considered in [6].…”
Section: Optimal Control and The Hamilton-jacobi-bellman Equation Formentioning
confidence: 99%
“…The policy restores the system to a previous, not necessarily AGAN, condition with certain probability. Similarly, Boussemart, Bickard, & Limnios (2001) considered a Markov chain that governs the system degradation, maintenance actions bring the system to a new state with certain probability, the new system state depends on the performed action. More details in this subject can be read in Section 1 .…”
Section: Maintenance In Hmmmentioning
confidence: 99%
“…Both of these reinforcement learning methods assume that every Markov chain induced by a policy is irreducible, which allows only a single recurrent class as with ergodic and unichain assumptions described earlier. The Lagrangian approach has also been applied to specific stochastic policy linear programming formulations relevant to aircraft maintenance problems where the asymptotic failure is to be kept below some small threshold (Boussemart & Limnios, 2004;Boussemart et al, 2002).…”
Section: Related Workmentioning
confidence: 99%
“…Steady-state planning has applications in several areas, such as deriving maintenance plans for various systems, including aircraft maintenance, where the asymptotic failure rate of components must be kept below some small threshold (Boussemart & Limnios, 2004;Boussemart, Limnios, & Fillion, 2002). Optimal routing problems for communication networks have also been proposed in which data throughput must be maximized subject to constraints on average delay and packet drop metrics (Lazar, 1983).…”
Section: Introductionmentioning
confidence: 99%