Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence 2018
DOI: 10.24963/ijcai.2018/337
|View full text |Cite
|
Sign up to set email alerts
|

Episodic Memory Deep Q-Networks

Abstract: Reinforcement learning (RL) algorithms have made huge progress in recent years by leveraging the power of deep neural networks (DNN). Despite the success, deep RL algorithms are known to be sample inefficient, often requiring many rounds of interaction with the environments to obtain satisfactory performance. Recently, episodic memory based RL has attracted attention due to its ability to latch on good actions quickly. In this paper, we present a simple yet effective biologically inspired RL algorithm called E… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
37
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 50 publications
(37 citation statements)
references
References 0 publications
0
37
0
Order By: Relevance
“…It learns faster than Double DQN or N-step DQN in the Atari game Pong, but it impairs the advantage of generalization of DQN and lacks the continuous and effective use of episodic memory. Recently, the author in [26] combines parametric module of DQN with nonparametric module of episodic control with the purpose of improving both sample efficiency as well as module generalization. This EMDQN method is better than DQN and surpasses both MFEC and NEC.…”
Section: B Incorporation Of Episodic Memorymentioning
confidence: 99%
See 2 more Smart Citations
“…It learns faster than Double DQN or N-step DQN in the Atari game Pong, but it impairs the advantage of generalization of DQN and lacks the continuous and effective use of episodic memory. Recently, the author in [26] combines parametric module of DQN with nonparametric module of episodic control with the purpose of improving both sample efficiency as well as module generalization. This EMDQN method is better than DQN and surpasses both MFEC and NEC.…”
Section: B Incorporation Of Episodic Memorymentioning
confidence: 99%
“…The parameters of EMDQN and DQN are all set the same as [26]. As for HE-EMDQN, the networks and basic hyperparameter settings are set as DQN.…”
Section: A Experimental Setupmentioning
confidence: 99%
See 1 more Smart Citation
“…One of such approaches is known as memory consolidation or system-level consolidation (McClelland et al, 1995): an episodic memory system maintains a subset of previously experienced sensorimotor data and replays them, along with the new samples, to the networks during the training. Episodic memory system has been integrated recently also in the deep learning systems, such as in Deep Q-Networks implementing deep reinforcement learning (RL; Lin et al, 2018).…”
Section: Introductionmentioning
confidence: 99%
“…ADQN can reduce estimation bias, but needs more networks, occupies more storage resources and affects calculating efficiency. Lin et al [34] combined episodic control with DQN, and proposed episodic memory deep Q-networks (EMDQN), which leverages episodic memory to supervise an agent during training. In EMDQN, it requires fewer interactions with the environment, generates better sample efficiency, and can also alleviate overestimation of DQN.…”
Section: Introductionmentioning
confidence: 99%