2016
DOI: 10.1609/icaps.v26i1.13773
|View full text |Cite
|
Sign up to set email alerts
|

From FOND to Robust Probabilistic Planning: Computing Compact Policies that Bypass Avoidable Deadends

Abstract: We address the class of probabilistic planning problems where the objective is to maximize the probability of reaching a prescribed goal. The complexity of probabilistic planning problems makes it difficult to compute high quality solutions for large instances, and existing algorithms either do not scale, or do so at the expense of the solution quality. We leverage core similarities between probabilistic and fully observable non-deterministic (FOND) planning to construct a sound, offline probabilistic planner,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(10 citation statements)
references
References 12 publications
0
10
0
Order By: Relevance
“…LAO* generalizes heuristic search to solving belief MDPs (Hansen and Zilberstein 2001). PO-PRP (Muise, Belle, and McIlraith 2014) and ProbPRP (Camacho, Muise, and McIlraith 2016) use a series of calls to classical planners to iteratively construct and refine a policy for planning in a partially observable environment. However, these two methods neither exploit clear preferences nor aim to provide any guarantee on the value of the computed policy.…”
Section: Related Workmentioning
confidence: 99%
“…LAO* generalizes heuristic search to solving belief MDPs (Hansen and Zilberstein 2001). PO-PRP (Muise, Belle, and McIlraith 2014) and ProbPRP (Camacho, Muise, and McIlraith 2016) use a series of calls to classical planners to iteratively construct and refine a policy for planning in a partially observable environment. However, these two methods neither exploit clear preferences nor aim to provide any guarantee on the value of the computed policy.…”
Section: Related Workmentioning
confidence: 99%
“…Far beyond the standard benchmarks in Table 1 (triangle-side length 20), VI on BS scales to side length 74 in both the original domain and the limited-budget version. For comparison, the hitherto best solver by far was Prob-PRP (Camacho et al 2016), which scales to side length 70 on the original domain, 6 and is optimal only for goal probability 1, i. e., in the presence of strong cyclic plans.…”
Section: Acyclic Planningmentioning
confidence: 99%
“…Kolobov et al (2012) and Teichteil (2012) consider objectives asking for the cheapest policy among those maximizing goal probability, also requiring FRET or VI. Other works addressing goal probability maximization (e. g. (Teichteil-Königsbuch, Kuter, and Infantes 2010;Camacho et al 2016)) do not aim at guaranteeing optimality. In summary, heuristic search for MaxProb is challenging, and has only been addressed by Kolobov et al (2011).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…For SSPs with dead ends, some research has focused only on finding policies that maximize the probability of reaching a goal (MAXPROB criterion) (Kolobov et al 2011;Teichteil-Königsbuch, Kuter, and Infantes 2010;Camacho, Muise, and McIlraith 2016), while other approaches work with two criteria: maximizing the probability of reaching a goal and minimizing the average accumulated costs of reaching a goal (Teichteil-Königsbuch 2012;Kolobov, Mausam, and Weld 2012;Trevizan, Teichteil-Königsbuch, and Thiébaux 2017).…”
Section: Introductionmentioning
confidence: 99%