Proceedings of the 33rd Annual ACM/IEEE Symposium on Logic in Computer Science 2018
DOI: 10.1145/3209108.3209185
|View full text |Cite
|
Sign up to set email alerts
|

Distribution-based objectives for Markov Decision Processes

Abstract: We consider distribution-based objectives for Markov Decision Processes (MDP). This class of objectives gives rise to an interesting trade-off between full and partial information. As in full observation, the strategy in the MDP can depend on the state of the system, but similar to partial information, the strategy needs to account for all the states at the same time.In this paper, we focus on two safety problems that arise naturally in this context, namely, existential and universal safety. Given an MDP A and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
11
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(11 citation statements)
references
References 28 publications
0
11
0
Order By: Relevance
“…The global state of the game is a distribution over the local states of the processes, and the specification describes which sequences of distributions are winning. The distributions can be discrete [AAGT12,CFO20] or continuous [KVAK10,AGV18]. The control may be applied uniformly, independently of the local state of each process, as in non-deterministic [BDGG17], and probabilistic automata [CFO20], or it may depend on the local history of states, as in Markov decision processes (MDPs) [AGV18,DMS19].…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations
“…The global state of the game is a distribution over the local states of the processes, and the specification describes which sequences of distributions are winning. The distributions can be discrete [AAGT12,CFO20] or continuous [KVAK10,AGV18]. The control may be applied uniformly, independently of the local state of each process, as in non-deterministic [BDGG17], and probabilistic automata [CFO20], or it may depend on the local history of states, as in Markov decision processes (MDPs) [AGV18,DMS19].…”
Section: Introductionmentioning
confidence: 99%
“…The distributions can be discrete [AAGT12,CFO20] or continuous [KVAK10,AGV18]. The control may be applied uniformly, independently of the local state of each process, as in non-deterministic [BDGG17], and probabilistic automata [CFO20], or it may depend on the local history of states, as in Markov decision processes (MDPs) [AGV18,DMS19]. In both cases imperfect information arises: either because the control is global, thus not aware of the local state of individual processes, or because the control is local, thus not aware of the global states on which the specification is defined.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…In the more recent distribution-based semantics, the outcome of a stochastic process is a sequence of distributions over states [3,19]. This alternative semantics has received some attention recently for theoretical analysis of probabilistic bisimulation [17] and is adequate to describe large populations of agents [14,10] with applications in system biology [19,1]. The behaviour of an agent is modeled as an MDP with some state space Q, and a large population of identical agents is described by a (continuous) distribution d : Q → [0, 1] that gives the fraction d(q) of agents in the population that are in each state q ∈ Q.…”
Section: Introductionmentioning
confidence: 99%