2015
DOI: 10.1016/j.artint.2015.05.008
|View full text |Cite
|
Sign up to set email alerts
|

Transferring knowledge as heuristics in reinforcement learning: A case-based approach

Abstract: The goal of this paper is to propose and analyse a transfer learning meta-algorithm that allows the implementation of distinct methods using heuristics to accelerate a Reinforcement Learning procedure in one domain (the target) that are obtained from another (simpler) domain (the source domain). This meta-algorithm works in three stages: first, it uses a Reinforcement Learning step to learn a task on the source domain, storing the knowledge thus obtained in a case base; second, it does an unsupervised mapping … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 59 publications
(27 citation statements)
references
References 32 publications
0
22
0
Order By: Relevance
“…Future work includes the addition of advanced techniques for boosting the learning process in robotics, e.g., by constructing a model (model-based RL), such as prioritized sweeping, by human interaction as teaching, and by transferring (Taylor & Stone 2009), (Barrett, Taylor & Stone 2010) and (Bianchi, Celiberto, Santos, Matsuura & de Mantaras 2015) from simulations to real robots. We also believe that these techniques will contribute to decrease the variance of the learning processes seen here, which sometimes have produced inconclusive results, especially in episodic tasks.…”
Section: Discussionmentioning
confidence: 99%
“…Future work includes the addition of advanced techniques for boosting the learning process in robotics, e.g., by constructing a model (model-based RL), such as prioritized sweeping, by human interaction as teaching, and by transferring (Taylor & Stone 2009), (Barrett, Taylor & Stone 2010) and (Bianchi, Celiberto, Santos, Matsuura & de Mantaras 2015) from simulations to real robots. We also believe that these techniques will contribute to decrease the variance of the learning processes seen here, which sometimes have produced inconclusive results, especially in episodic tasks.…”
Section: Discussionmentioning
confidence: 99%
“…In the TBO, 50% of bees with nectar amounts that rank in the half top of all bees are designated as worker, while the others are scout. On the basis of the ε-Greedy rule [31], the scouts' actions are based on the proportion of Q-value in the current status. As for the controlled variable x i , one behavior of each scout is chosen as…”
Section: Action Policymentioning
confidence: 99%
“…Also, reinforcement learning (RL) can be accelerated by knowledge conversion [29], and agents learn new tasks faster and interact less with the environment. As a consequence, knowledge transfer reinforcement learning (KTRL) has been developed [30] through combining AI and behavior psychology [31] and is divided into behavior shift and information shift.…”
Section: Introductionmentioning
confidence: 99%
“…In this sense, the learned pheromone distribution could help into defining a function to generate new trajectories and therefore improve the results of BO. The learning of heuristic functions is also important towards transfer learning [see the contribution of Weiss et al (2016) for a survey of transfer learning applied on several domains] in the reinforcement learning domain, as also shown by Bianchi et al (2015). With little modification, the heuristic function learned with AP-HARL could be simply used as a model for a particular task or problem and then re-used in a related problem.…”
Section: Related Workmentioning
confidence: 99%