DOI: 10.1007/978-3-540-74958-5_70
|View full text |Cite
|
Sign up to set email alerts
|

Transfer Learning in Reinforcement Learning Problems Through Partial Policy Recycling

Abstract: Abstract. In this paper we investigate the relation between transfer learning in reinforcement learning with function approximation and supervised learning with concept drift. We present a new incremental relational regression tree algorithm that is capable of dealing with concept drift through tree restructuring and show that it enables a reinforcement learner, more precisely a Q-learner, to transfer knowledge from one task to another by recycling those parts of the generalized Q-function that still hold inte… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
22
0

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 47 publications
(22 citation statements)
references
References 10 publications
0
22
0
Order By: Relevance
“…Ramon et al (2007) extend the TG algorithm with tree restructuring operations based on ideas of the incremental tree induction (ITI) algorithm by Utgoff (1997) and concepts from theory revision in ILP. New operations in this new algorithm TGR include the pruning of a leaf or a whole subtree, and the revision of an internal node.…”
Section: Logical Abstractionsmentioning
confidence: 99%
“…Ramon et al (2007) extend the TG algorithm with tree restructuring operations based on ideas of the incremental tree induction (ITI) algorithm by Utgoff (1997) and concepts from theory revision in ILP. New operations in this new algorithm TGR include the pruning of a leaf or a whole subtree, and the revision of an internal node.…”
Section: Logical Abstractionsmentioning
confidence: 99%
“…Policy transfer used in company with various RL methods has been applied to boost solution quality in various single-agent tasks including pole-balancing (Ammar et al, 2012), gameplaying (Ramon et al, 2007), robot navigation, as well as multiagent tasks including predator-prey (Boutsioukis et al, 2012). For such single and multiagent tasks, policy transfer is typically done within the same task domain for varying task complexity (Torrey and Shavlik, 2009) and less frequently between different task domains (Bou-Ammar et al, 2015).…”
Section: Evolutionary Policy Transfermentioning
confidence: 99%
“…The result is a new covering algorithm that utilises "Transfer Knowledge" approach to fill in missing attribute values specifically the target attribute in the training dataset before mining kicks in. The "Transfer Knowledge" is basically building up a knowledge base based on learning curves via agents from different environments and from previous learning experience and then using this knowledge base to fill in incomplete data examples (Ramon et al, 2007). Experimental results against eight datasets revealed that the improved covering algorithm consistently produced competitive classifiers when compared with other RI algorithms such as RULES-IS and PRISM.…”
Section: Prism Consmentioning
confidence: 99%