2018
DOI: 10.1017/s0269888918000280
|View full text |Cite
|
Sign up to set email alerts
|

Q-Table compression for reinforcement learning

Abstract: Reinforcement learning (RL) algorithms are often used to compute agents capable of acting in environments without prior knowledge of the environment dynamics. However, these algorithms struggle to converge in environments with large branching factors and their large resulting state-spaces. In this work, we develop an approach to compress the number of entries in a Q-value table using a deep auto-encoder. We develop a set of techniques to mitigate the large branching factor problem. We present the application o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 20 publications
0
7
0
Order By: Relevance
“…Each robot observes the environment locally and takes the action without relying on a centralized controller for synchronization between robots thus not having any direct information about other robot's intensions. However, it is a good practice to hugely accelerate the learning process, alleviate branching factor problem [35] and gain more confidence on learned policies visiting most of the states in state space several times by sharing the observations of all robots only in learning phase without harming the decentralization concept in exploitation phase. We should bear in mind that sharing the observations is not a necessary task for our method.…”
Section: B) Contributionmentioning
confidence: 99%
See 2 more Smart Citations
“…Each robot observes the environment locally and takes the action without relying on a centralized controller for synchronization between robots thus not having any direct information about other robot's intensions. However, it is a good practice to hugely accelerate the learning process, alleviate branching factor problem [35] and gain more confidence on learned policies visiting most of the states in state space several times by sharing the observations of all robots only in learning phase without harming the decentralization concept in exploitation phase. We should bear in mind that sharing the observations is not a necessary task for our method.…”
Section: B) Contributionmentioning
confidence: 99%
“…However, our method does not rely on coordination between agents and despite QMIX, learning phase and running phase both are done in a decentralized fashion. There are some other researches that try to compress the state space and reduce dimensionality [35], [45]. These researches utilize a method named deep auto-encoders for compressing the state space utilizing neural networks, mainly with the aim of gaining generalization ability rather than memory usage reduction.…”
Section: C) Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Artificial neural networks (ANNs) are a type of machine learning model made up of numerous nodes grouped in layers that compute an output depending on node activation mediated by weights in the connections between them. ANNs are capable of solving a variety of machine learning tasks, including classification, regression, and dimensionality reduction [44].…”
Section: Ae-based Deep Clusteringmentioning
confidence: 99%
“…The fourth paper Q-table compression for reinforcement learning by Rosa Amado and Meneguzzi (2018) proposes a method to reduce the number of entries in a Q-value table by using a deep autoencoder. Multi-agent reinforcement learning where agents share experience updates is also applied to mitigate the large branching factors which are present when controlling teams of units in real-time strategy (RTS) games.…”
Section: Contents Of the Special Issuementioning
confidence: 99%