2011
DOI: 10.1613/jair.3231
|View full text |Cite
|
Sign up to set email alerts
|

Scaling up Heuristic Planning with Relational Decision Trees

Abstract: Current evaluation functions for heuristic planning are expensive to compute. In numerous planning problems these functions provide good guidance to the solution, so they are worth the expense. However, when evaluation functions are misguiding or when planning problems are large enough, lots of node evaluations must be computed, which severely limits the scalability of heuristic planners. In this paper, we present a novel solution for reducing node evaluations in heuristic planning based on machine learning. P… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
40
0

Year Published

2013
2013
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 28 publications
(40 citation statements)
references
References 28 publications
0
40
0
Order By: Relevance
“…Accordingly, typed sequences can be complemented with other techniques that handle node evaluation issues. As future work we are planning to use CBR node recommendations to validate the action policy of ROLLER [9]. The new algorithm will propose a set of candidate actions to be applied using the relational decision trees, and then will select the actions that replay a typed sequence stored in the case base.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Accordingly, typed sequences can be complemented with other techniques that handle node evaluation issues. As future work we are planning to use CBR node recommendations to validate the action policy of ROLLER [9]. The new algorithm will propose a set of candidate actions to be applied using the relational decision trees, and then will select the actions that replay a typed sequence stored in the case base.…”
Section: Discussionmentioning
confidence: 99%
“…It performed poorly in the competition because the algorithm (not presented in this work) strongly depends on the quality of the relaxed plan. Other systems such as OBTUSEWEDGE [27], and ROLLER [9], performed better. These systems learn a generalized policy [19], which is a function that maps a metastate to the action that should be applied.…”
Section: Related Workmentioning
confidence: 99%
“…For example, TLPlan (Bacchus & Kabanza, 2000) and TALPlan (Kvarnström & Doherty, 2000) make use of control rules to prune the search space while hierarchical task network planners such as O-Plan (Tate, Drabble, & Kirby, 1994) and SHOP2 (Nau et al, 2003) use domain-specific knowledge for decomposing tasks into subtasks. It is also worth mentioning the work done by the planning and learning community who are looking to learn search control knowledge for planning automatically (de la Rosa, Jiménez, Fuentetaja, & Borrajo, 2011;Krajanský, Hoffmann, Buffet, & Fern, 2014) (see also the learning track of the international planning competition). The justification for the restriction of search control knowledge is that human planners are able to solve problems without being given such domain-specific rules.…”
Section: Ai Planning Vs Constraint Programmingmentioning
confidence: 99%
“…6 Symmetry breaking reuses information about one object to its symmetric "twin" in such a way that "bad" states of one object can be avoided for its symmetric "twin." 10 With growing interest in extracting DCK automatically, emphasis was given on exploiting machine learning techniques that can acquire useful DCK, usually, by analyzing "training plans," which are solutions of simple planning tasks. 4 Another way how performance of planning engines can be improved is by gathering domain control knowledge (DCK), ie, additional knowledge about planning tasks indicating how solution plans would look like.…”
Section: Introductionmentioning
confidence: 99%
“…DCK can be expressed, for instance, in the form of control rules, 8 temporal logic formulas, 9 or decision trees. 10 With growing interest in extracting DCK automatically, emphasis was given on exploiting machine learning techniques that can acquire useful DCK, usually, by analyzing "training plans," which are solutions of simple planning tasks. This motivated the foundation of the learning track in the IPC, which has been organized since 2008.…”
Section: Introductionmentioning
confidence: 99%