2022
DOI: 10.1609/icaps.v32i1.19845
|View full text |Cite
|
Sign up to set email alerts
|

Neural Network Heuristic Functions for Classical Planning: Bootstrapping and Comparison to Other Methods

Abstract: How can we train neural network (NN) heuristic functions for classical planning, using only states as the NN input? Prior work addressed this question by (a) per-instance imitation learning and/or (b) per-domain learning. The former limits the approach to instances small enough for training data generation, the latter to domains where the necessary knowledge generalizes across instances. Here we explore three methods for (a) that make training data generation scalable through bootstrapping and approximate val… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(19 citation statements)
references
References 15 publications
0
10
0
Order By: Relevance
“…The coverage of a planner is defined as the percent of initial states for which a solution path is found within the given planning budget. Ferber et al (2021) report observing that in general the coverage superiority between the different NN heuristics tested did not vary over time. That is, the planning time used and the relative coverage superiority between the algorithms were not correlated.…”
Section: Experimental Studymentioning
confidence: 92%
See 3 more Smart Citations
“…The coverage of a planner is defined as the percent of initial states for which a solution path is found within the given planning budget. Ferber et al (2021) report observing that in general the coverage superiority between the different NN heuristics tested did not vary over time. That is, the planning time used and the relative coverage superiority between the algorithms were not correlated.…”
Section: Experimental Studymentioning
confidence: 92%
“…The baseline option for performing the rollout, and the method used by Ferber et al (2021), is to randomly select actions a for which applying the regression operator is valid. In addition to testing RSL using random action selection we instantiate a version of RSL we name Novelty guided Regression based Learning (N-RSL) that aims to increase the structural diversity of operators selected in its regression.…”
Section: Extracting State Sets Through Regressionmentioning
confidence: 99%
See 2 more Smart Citations
“…Successes include the AlphaGo series (Silver et al 2016(Silver et al , 2017(Silver et al , 2018, as well as heuristic search for single-agent games such as Rubik's Cube (Agostinelli et al 2019). Given the prominence of heuristic search in AI Planning (Hoffmann and Nebel 2001;Helmert and Domshlak 2009;Richter and Westphal 2010;Helmert et al 2014;Domshlak, Hoffmann, and Katz 2015), training NNs as heuristic functions is highly promising, and is actively pursued (Toyer et al 2018;Garg, Bajpai, and Mausam 2019;Ferber, Helmert, and Hoffmann 2020;Shen, Trevizan, and Thiébaux 2020;Rivlin, Hazan, and Karpas 2020;Yu, Kuroiwa, and Fukunaga 2020;Karia and Srivastava 2021;Ferber et al 2022). We contribute a new angle to this line of research, honing in on NN prediction confidence.…”
Section: Introductionmentioning
confidence: 99%