Robot Learning 1993
DOI: 10.1007/978-1-4615-3184-5_3
|View full text |Cite
|
Sign up to set email alerts
|

Learning Multiple Goal Behavior via Task Decomposition and Dynamic Policy Merging

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
29
0

Year Published

1996
1996
2011
2011

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 71 publications
(29 citation statements)
references
References 0 publications
0
29
0
Order By: Relevance
“…Each module has its own Q matrix representing its partial knowledge of the world state s i . An agent-level module selection or action selection policy chooses an action from the module preferences, such as the Greatest Mass (GM) strategy [49,75]:…”
Section: Modular Reinforcement Learningmentioning
confidence: 99%
“…Each module has its own Q matrix representing its partial knowledge of the world state s i . An agent-level module selection or action selection policy chooses an action from the module preferences, such as the Greatest Mass (GM) strategy [49,75]:…”
Section: Modular Reinforcement Learningmentioning
confidence: 99%
“…Many modular architectures have been proposed so far (for example [7], [8], [9]). Each module is responsible for learning to achieve a single Fig.2 shows a sketch of such a modular learning system.…”
Section: B Modular Learning Systemmentioning
confidence: 99%
“…To remedy the problem of combinatorial explosion in multiagent reinforcement learning, Whitehead [39] proposed an architecture called modular Q-learning, which decomposes the whole problem space into smaller subproblem spaces and distributes them among multiple modules. This way, the goals of multiple-goal problems are decomposed into subgoals, which are then distributed among multiple modules.…”
Section: B Multiagent Learning Fuzzy Modular Approach With Internal mentioning
confidence: 99%