2018 IEEE Conference on Computational Intelligence and Games (CIG) 2018
DOI: 10.1109/cig.2018.8490400
|View full text |Cite
|
Sign up to set email alerts
|

Anxious Learning in Real-Time Heuristic Search

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
13
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(14 citation statements)
references
References 19 publications
1
13
0
Order By: Relevance
“…The second way of thinking about the performance degradation lies with the notion of heuristic depressions (Ishida 1997) which has been subject of recent research (Hernández and Baier 2014;Sturtevant and Bulitko 2016;Bulitko and Sampley 2016). Viewed in terms of heuristic depressions, the degradation of LRTA* with self-knowledge can be viewed as follows.…”
Section: Counter Examplementioning
confidence: 99%
“…The second way of thinking about the performance degradation lies with the notion of heuristic depressions (Ishida 1997) which has been subject of recent research (Hernández and Baier 2014;Sturtevant and Bulitko 2016;Bulitko and Sampley 2016). Viewed in terms of heuristic depressions, the degradation of LRTA* with self-knowledge can be viewed as follows.…”
Section: Counter Examplementioning
confidence: 99%
“…We assume that: (A1) action costs are greater than 0, (A2) for every state, there is a goal reachable from it, (A3) all initial beliefs have finite expected value, (A4) the state space is finite. Our proof follows the style of Korf's (1990) proof for RTA* and Bulitko and Sampley's (2016) proof for Weighted Lateral LRTA* (wbLRTA*): we first prove that incompleteness implies that there must exist a subset of states within which Nancy circulates forever. Then we prove that there cannot exist such a set due to the updates made by the learning rule of Nancy.…”
Section: Theoretical Analysismentioning
confidence: 99%
“…LRTA* originally updated the heuristic value of a state based on the surrounding state with the minimum heuristic estimate to the goal. Different heuristic learning rules have also been researched for alternatives to the original methods used in LRTA* (Bulitko 2016d;2016a). Weighting the learning of the agent has been explored and shown to help speed up converging to the correct heuristic value (Shimbo and Ishida 2003;Rivera, Baier, and Hernndez 2015).…”
Section: Introduction and Related Workmentioning
confidence: 99%