2012
DOI: 10.1007/978-3-642-27645-3_9
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical Approaches

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(14 citation statements)
references
References 29 publications
0
14
0
Order By: Relevance
“…This work highlights the fact that many sequential decision problems contain repeated sets of actions that are related to solving the same subgoals. At a high level, many of the computational approaches for learning which subgoals are useful for a given task involve identifying commonly traversed task states that serve as bottlenecks between large sets of similar states (Hengst, 2012; Tomov, Yagati, Kumar, Yang, & Gershman, 2020). For example, a doorway between two rooms represents a bottleneck linking any state in the first room to any state in the second room.…”
Section: Task Representationsmentioning
confidence: 99%
“…This work highlights the fact that many sequential decision problems contain repeated sets of actions that are related to solving the same subgoals. At a high level, many of the computational approaches for learning which subgoals are useful for a given task involve identifying commonly traversed task states that serve as bottlenecks between large sets of similar states (Hengst, 2012; Tomov, Yagati, Kumar, Yang, & Gershman, 2020). For example, a doorway between two rooms represents a bottleneck linking any state in the first room to any state in the second room.…”
Section: Task Representationsmentioning
confidence: 99%
“…The famous 'divide-and-conquer' strategy has been practiced for RL research for a few decades, with a number of studies which show that dividing the problem into sub-problems and making abstractions based on them can significantly improve learning performance (Dietterich, 2000;Hengst, 2012). However, it is not always straightforward to devise a meaningful partitioning scheme.…”
Section: Automatic Abstraction In Reinforcement Learningmentioning
confidence: 99%
“…Our DMSSPs model share elements with previously studied MDP models: arbitrarily modulated transition functions [8], stochastic shortest paths with online information [9], and factored hybrid-space MDPs [10]. Our HSP algorithm uses ideas from heuristic search [11,12] and search-based planning for multi-step tasks [13,14], approximate dynamic programming [15,6], hierarchical planning for solving large MDPs [16,17,18], and interleaved planning and execution [19,20]. A body of relevant previous work incorporates heuristic search and classical AI techniques in algorithms for solving MDPs [21,22,23].…”
Section: Related Work Overviewmentioning
confidence: 99%