2020
DOI: 10.3389/frobt.2020.00097
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Channel Interactive Reinforcement Learning for Sequential Tasks

Abstract: The ability to learn new tasks by sequencing already known skills is an important requirement for future robots. Reinforcement learning is a powerful tool for this as it allows for a robot to learn and improve on how to combine skills for sequential tasks. However, in real robotic applications, the cost of sample collection and exploration prevent the application of reinforcement learning for a variety of tasks. To overcome these limitations, human input during reinforcement can be beneficial to speed up learn… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(9 citation statements)
references
References 31 publications
0
9
0
Order By: Relevance
“…Such explicit rewards are used incorporate domain knowledge into RL [15], [16]. Alternatively, teacher intervention can serve as implicit negative rewards, indicating that actions are undesirable [17], [18].…”
Section: A Learning From Ne Feedbackmentioning
confidence: 99%
“…Such explicit rewards are used incorporate domain knowledge into RL [15], [16]. Alternatively, teacher intervention can serve as implicit negative rewards, indicating that actions are undesirable [17], [18].…”
Section: A Learning From Ne Feedbackmentioning
confidence: 99%
“…Firstly, each type of human-robot interaction is more adequate to capture different stages of the learning process. Secondly, different types of information are used during the learning process, which can achieve a faster or more consistent learning rate (Koert et al, 2020).…”
Section: Interactions With Multiple Interfacesmentioning
confidence: 99%
“…This channel allows the robot to communicate the necessity of corrections or input, as in the case of robot-gated or active learning methods, in which the user is asked for feedback only in specific cases which improve the learning process (e.g., Sadigh et al, 2017;Cui and Niekum, 2018;Brown et al, 2018;Biyik and Sadigh, 2018;Biyik et al, 2020;Hoque et al, 2022;Franzese et al, 2021a). Additionally, Li et al (2016) and Koert et al (2020) show that providing the human with uncertainty and performance information about the robot's actions can improve the teacher's feedback quality, improving the learning process, i.e., the teacher concurrently learning how to teach the robot. Thus, understanding how and what information to communicate to the human can be a key to enabling non-experts to interactively teach robots.…”
Section: Robot-to-human Interfacesmentioning
confidence: 99%
“…Additionally, these interactive methods enable users to transfer their knowledge with other modalities of interaction, not only explicitly showing what the agent should do, but also guiding with evaluative feedback from the teacher, i.e., reinforcements (rewards or punishments), or also by comparing the performance of different agents/policies with learning from preferences or rankings. The use of evaluative feedback bridges the worlds of RL and IL [13][14][15][16].…”
Section: Introductionmentioning
confidence: 99%