2022
DOI: 10.7494/csci.2022.23.1.4139
|View full text |Cite
|
Sign up to set email alerts
|

Improving modified policy iteration for probabilistic model checking

Abstract: Value iteration, policy iteration and their modified versions are well-known algorithms for probabilistic model checking of Markov Decision Processes. One the challenge of these methods is that they are time-consuming in most cases. Several techniques have been proposed to improve the performance of iterative methods for probabilistic model checking. However, the running time of these techniques depends on the graphical structure of the model and in some cases their performance is worse than the performance of… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
1
1
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 43 publications
0
4
0
Order By: Relevance
“…The modified policy iteration [1] method updates the policies after a fixed number of iterations (100 iterations, for example) rather than satisfying the convergence criterion. It usually causes faster convergence to the estimations of values [29,32]. More details of modified policy iteration are available in [1,19,31].…”
Section: Modified Policy Iterationmentioning
confidence: 99%
See 3 more Smart Citations
“…The modified policy iteration [1] method updates the policies after a fixed number of iterations (100 iterations, for example) rather than satisfying the convergence criterion. It usually causes faster convergence to the estimations of values [29,32]. More details of modified policy iteration are available in [1,19,31].…”
Section: Modified Policy Iterationmentioning
confidence: 99%
“…Although modified policy iteration is usually faster than the standard policy iteration (Algorithm 1), the number of iterations of the Gauss-Seidel method influences its running time. In [29] a dynamic method has been proposed for determining the number of iterations for every quotient DTMCs.…”
Section: Modified Policy Iterationmentioning
confidence: 99%
See 2 more Smart Citations