2020
DOI: 10.1007/s10489-020-01679-3
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement learning with convolutional reservoir computing

Abstract: Recently, reinforcement learning models have achieved great success, mastering complex tasks such as Go and other games with higher scores than human players. Many of these models store considerable data on the tasks and achieve high performance by extracting visual and time-series features using convolutional neural networks (CNNs) and recurrent neural networks, respectively. However, these networks have very high computational costs because they need to be trained by repeatedly using the stored data. In this… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 20 publications
(8 citation statements)
references
References 26 publications
0
8
0
Order By: Relevance
“…Notwithstanding these issues, we are optimistic that the ideas behind MARC can have far-reaching application beyond the focus of time series prediction presented in this work. Firstly, it may be the case that RCs trained for other tasks (such as control [45,46], signal classification [47,48], and anomaly detection [49,50]) or through reinforcement learning [51] can be similarly compressed with autoencoders so that new, related systems can be controlled or classified with greatly reduced training data requirements. Secondly, we have focused here on RCs-both because of our initial focus on time series prediction and due to attractive numerical properties-but the general scheme of Sec.…”
Section: Discussionmentioning
confidence: 99%
“…Notwithstanding these issues, we are optimistic that the ideas behind MARC can have far-reaching application beyond the focus of time series prediction presented in this work. Firstly, it may be the case that RCs trained for other tasks (such as control [45,46], signal classification [47,48], and anomaly detection [49,50]) or through reinforcement learning [51] can be similarly compressed with autoencoders so that new, related systems can be controlled or classified with greatly reduced training data requirements. Secondly, we have focused here on RCs-both because of our initial focus on time series prediction and due to attractive numerical properties-but the general scheme of Sec.…”
Section: Discussionmentioning
confidence: 99%
“…Tong and Tanaka [49] took full advantage of an untrained CNN to transform raw image data into a set of small feature maps as a preprocessing step of RC and achieved a high classification accuracy with a much smaller number of trainable parameters compared with Schaetti's work. Then, a novel practical approach, called RL with a convolutional reservoir computing (RCRC) model [5] was proposed. A fixed random-weight CNN, used for extracting image feature maps, combined with an RC model, employed for extracting the time series features, is adopted for the evolution strategy of RL, which succeeded in decreasing computational cost to a large degree.…”
Section: Related Workmentioning
confidence: 99%
“…Reservoir computing has attracted significant attention in various research fields because it is capable of fast learning that results in reduced computational/training costs compared to other recurrent neural networks [19,20]. Reservoir computing is a computation framework used for information processing and supervised learning [21,22].…”
Section: Introductionmentioning
confidence: 99%