2013 12th International Conference on Machine Learning and Applications 2013
DOI: 10.1109/icmla.2013.67
|View full text |Cite
|
Sign up to set email alerts
|

A Temporal Difference GNG-Based Algorithm That Can Learn to Control in Reinforcement Learning Environments

Abstract: This paper proposes a new reinforcement learning algorithm called TD-GNG that uses the Growing Neural Gas (GNG) network to deal with environments of large domains. The proposed algorithm is capable to reduce the dimensionality of the problem by aggregating similar states. In experimental comparison against tile-coding in mountain car and puddle world, the TD-GNG showed an increase in the generalization without loosing quality in the policy obtained.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2015
2015
2016
2016

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 6 publications
0
2
0
Order By: Relevance
“…Due to high-dimensional and real-valued state spaces, it is usually not feasible to learn a suitable selection policy for each state individually. Aggregation algorithms, e.g., [9,10,6,13,1], dynamically partition the state space of a reinforcement learning problem into disjunct macro states do deal with this problem. Typically, these algorithms start with a coarse-grained partitioning of the state space and refine the state space based on various conditions.…”
Section: Aggregated State Spacesmentioning
confidence: 99%
See 1 more Smart Citation
“…Due to high-dimensional and real-valued state spaces, it is usually not feasible to learn a suitable selection policy for each state individually. Aggregation algorithms, e.g., [9,10,6,13,1], dynamically partition the state space of a reinforcement learning problem into disjunct macro states do deal with this problem. Typically, these algorithms start with a coarse-grained partitioning of the state space and refine the state space based on various conditions.…”
Section: Aggregated State Spacesmentioning
confidence: 99%
“…Another group of aggregation algorithms uses the idea of the nearest neighbor vector quantization to create macro states, e.g., [6,13]. These algorithms maintain a codebook CB ⊆ S containing specific states that are called codewords.…”
Section: Aggregated State Spacesmentioning
confidence: 99%