2004
DOI: 10.1007/978-3-540-30109-7_17
|View full text |Cite
|
Sign up to set email alerts
|

Learning Goal Hierarchies from Structured Observations and Expert Annotations

Abstract: Abstract. We describe a framework for generating agent programs that model expert task performance in complex dynamic domains, using expert behavior observations and goal annotations as the primary source. We map the problem of learning an agent program on to multiple learning problems that can be represented in a "supervised concept learning" setting. The acquired procedural knowledge is partitioned into a hierarchy of goals and it is represented with first order rules. Using an inductive logic programming (I… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
20
0

Year Published

2007
2007
2015
2015

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(20 citation statements)
references
References 17 publications
0
20
0
Order By: Relevance
“…The framework is based on time points and used only the successor temporal relation. Könik and Laird (2006) proposed a learning by observation framework to learn an agent program that mimics a human expert's behaviour in domains such as games. The learned concepts are used to generate behaviour rather than classification.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The framework is based on time points and used only the successor temporal relation. Könik and Laird (2006) proposed a learning by observation framework to learn an agent program that mimics a human expert's behaviour in domains such as games. The learned concepts are used to generate behaviour rather than classification.…”
Section: Related Workmentioning
confidence: 99%
“…The important aspect to note for the above review is that most of the work in this area has been done on either artificial or simulated data (Moyle & Muggleton, 1997;Könik & Laird, 2006) or very simple real world data (Fern et al, 2002;Needham et al, 2005) that involves few objects, the events are of short duration and all the objects in the scene are involved in the events. In our case, the tracked data from videos is very large and at the same time more complex and noisy and contains more objects.…”
Section: Related Workmentioning
confidence: 99%
“…Human demonstrations have also received some attention to speed up reinforcement learning (Schaal 1996), and as a way of automatically acquiring planning knowledge (Hogg, Muñoz-Avila, and Kuter 2008), among others. Könik and Laird present a Relational Learning from Observation technique (Könik and Laird 2006) able to learn how to decompose a goal into subgoals, based on observing annotated expert traces. Könik and Laird's technique uses relational machine learning techniques to learn how to decompose goals, and the output is a collection of rules, thus showing an approach to learning planning knowledge from demonstrations.…”
Section: Related Workmentioning
confidence: 99%
“…A number of systems (e.g, van Lent & Laird, 1999;Wang, 1995;Konik & Laird, 2006) have also been developed to learn procedural rules or plan operators from observations of expert behavior. Wang's OBSERVER (Wang, 1995) learns STRIPS style operators; van Lent's KnoMic (van Lent & Laird, 1999) learns production rules for the Soar agent architecture and Konik's system (Konik & Laird, 2006) creates first order logic rules that are later converted into Soar productions.…”
Section: Learning By Observationmentioning
confidence: 99%
“…Wang's OBSERVER (Wang, 1995) learns STRIPS style operators; van Lent's KnoMic (van Lent & Laird, 1999) learns production rules for the Soar agent architecture and Konik's system (Konik & Laird, 2006) creates first order logic rules that are later converted into Soar productions. All three systems use similar behavior traces as our approach, although Wang's OBSERVER works only with primitive actions so there is no notion of non-atomic goals and thus no need to annotate them in the behavior traces.…”
Section: Learning By Observationmentioning
confidence: 99%