2021 20th International Conference on Advanced Robotics (ICAR) 2021
DOI: 10.1109/icar53236.2021.9659344
|View full text |Cite
|
Sign up to set email alerts
|

Towards Hierarchical Task Decomposition using Deep Reinforcement Learning for Pick and Place Subtasks

Abstract: Deep Reinforcement Learning (DRL) is emerging as a promising approach to generate adaptive behaviors for robotic platforms. However, a major drawback of using DRL is the data-hungry training regime that requires millions of trial and error attempts, which is impractical when running experiments on robotic systems. To address this issue, we propose a multi-subtask reinforcement learning method where complex tasks are decomposed manually into low-level subtasks by leveraging human domain knowledge. These subtask… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
26
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 25 publications
(26 citation statements)
references
References 24 publications
(29 reference statements)
0
26
0
Order By: Relevance
“…Hierarchical reinforcement learning HRL [13,14] divides agents' tasks into sub-tasks to be learned by different agents. This simplifies the problem to be solved by each agent, making their behaviour easier to interpret and thereby making them easier to characterize.…”
Section: Agent Characterizationmentioning
confidence: 99%
“…Hierarchical reinforcement learning HRL [13,14] divides agents' tasks into sub-tasks to be learned by different agents. This simplifies the problem to be solved by each agent, making their behaviour easier to interpret and thereby making them easier to characterize.…”
Section: Agent Characterizationmentioning
confidence: 99%
“…As a remedy the broad area of Hierarchical Reinforcement Learning (HRL) attempts to decompose RL problems into multiple levels of abstraction-temporal, spatial, or otherwise. Many works deploy separate policies over different time horizons and action spaces (Barto & Mahadevan, 2003;Levy et al, 2017;Yang et al, 2020;Marzari et al, 2021). Temporal abstraction in planning can be traced back at least to Sutton et al (1999), where the options were introduced to refer to lower level policies.…”
Section: Related Workmentioning
confidence: 99%
“…Larger problems are solved by choreographing these subtasks through an orchestration agent that learns the highlevel dynamics of its environment [25]. HRL has, for instance, been used to control a robotic arm: while low-level agents learned simple tasks such as moving forward / backward or picking up / placing down, an orchestration agent learned to retrieve objects on a surface by choreographing these tasks [26,27]. The agents were not only efficient at learning, but their policies were more easily interpreted by human experts.…”
Section: Background and Related Workmentioning
confidence: 99%