2021
DOI: 10.1088/2634-4386/ac1a64
|View full text |Cite
|
Sign up to set email alerts
|

A dual-memory architecture for reinforcement learning on neuromorphic platforms

Abstract: Reinforcement learning (RL) is a foundation of learning in biological systems and provides a framework to address numerous challenges with real-world artificial intelligence applications. Efficient implementations of RL techniques could allow for agents deployed in edge-use cases to gain novel abilities, such as improved navigation, understanding complex situations and critical decision making. Toward this goal, we describe a flexible architecture to carry out RL on neuromorphic platforms. This architecture wa… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 35 publications
(32 reference statements)
0
3
0
Order By: Relevance
“…Equally, prioritized stochastic memory management (PSMM) [20], combined experience replay (CER) [37], selective experience replay (SER) [17], and episodic memory control (EMC) [40] use experience retention www.ijacsa.thesai.org strategies (memory management strategies). In contrast, some replay strategies focus on the structure of the replay memory instead of the content [35], [42], [43]. ERO has proven superior among prioritized selection algorithms, owing to its easy adaptation and generalization to multiple environments [23].…”
Section: Prioritized Sequence Experience Replay (Pser) [39] mentioning
confidence: 99%
“…Equally, prioritized stochastic memory management (PSMM) [20], combined experience replay (CER) [37], selective experience replay (SER) [17], and episodic memory control (EMC) [40] use experience retention www.ijacsa.thesai.org strategies (memory management strategies). In contrast, some replay strategies focus on the structure of the replay memory instead of the content [35], [42], [43]. ERO has proven superior among prioritized selection algorithms, owing to its easy adaptation and generalization to multiple environments [23].…”
Section: Prioritized Sequence Experience Replay (Pser) [39] mentioning
confidence: 99%
“…Just as multiple neural networks are used in some RL algorithms to enhance training stability [8,11,15,26,30,31], recent works in ER are exploring the use of a dual-memory architecture. The dualism may come in the form of long and short memory [32] or main and cache [33]. It may also be differentiated based on the sources of the replay data or the ratio of selection from the dual memory.…”
Section: Experience Selection Strategies and Algorithmsmentioning
confidence: 99%
“…It may also be differentiated based on the sources of the replay data or the ratio of selection from the dual memory. Olin-Ammentorp et al [32] rely on the complementary learning system of the human brain (interaction between the cortical and hippocampal networks) to design a dual memory (short-term and long-term) replay architecture. However, their design was implemented in a discrete state-action space.…”
Section: Experience Selection Strategies and Algorithmsmentioning
confidence: 99%