2021
DOI: 10.1145/3453160
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical Reinforcement Learning

Abstract: Hierarchical Reinforcement Learning (HRL) enables autonomous decomposition of challenging long-horizon decision-making tasks into simpler subtasks. During the past years, the landscape of HRL research has grown profoundly, resulting in copious approaches. A comprehensive overview of this vast landscape is necessary to study HRL in an organized manner. We provide a survey of the diverse HRL approaches concerning the challenges of learning hierarchical policies, subtask discovery, transfer learning, and multi-ag… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
90
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 207 publications
(90 citation statements)
references
References 31 publications
0
90
0
Order By: Relevance
“…Hierarchical RL (Pateria et al, 2022) is a way to learn, plan, and represent knowledge with temporal abstraction at multiple levels, with a long history, e.g., options (Sutton et al, 1999). Hierarchical RL is an approach for issues of sparse rewards and/or long horizons, with exploration in the space of high-level goals.…”
Section: More Challengesmentioning
confidence: 99%
“…Hierarchical RL (Pateria et al, 2022) is a way to learn, plan, and represent knowledge with temporal abstraction at multiple levels, with a long history, e.g., options (Sutton et al, 1999). Hierarchical RL is an approach for issues of sparse rewards and/or long horizons, with exploration in the space of high-level goals.…”
Section: More Challengesmentioning
confidence: 99%
“…We propose a two level Feudal Reinforcement Learning agent shown in Figure 2. Feudal Reinforcement Learning (Feudal RL) is a type of Hierarchical Reinforcement Learning where a high-level controller sets a subtask that is executed by a lower-level controller [Dayan and Hinton, 1992;Pateria et al, 2021]. The state space S = S h × S o consists of the positions and velocities of the agent's joints as well as the position and velocity of the projectile ball.…”
Section: R(s) =mentioning
confidence: 99%
“…Different from previous work on HRL [21,31] and symbolic planning [28,15], in this work we use ILP to learn a symbolic state transition model consisting of learned clauses. The ILP method adopted is a refinement-based ∂ILP which integrates refinement operation [5] and ∂ILP [10] together.…”
Section: Learning Symbolic Transition Model Via Ilpmentioning
confidence: 99%