2023
DOI: 10.3390/app13042034
|View full text |Cite
|
Sign up to set email alerts
|

Experience Replay Optimisation via ATSC and TSC for Performance Stability in Deep RL

Abstract: Catastrophic forgetting is a significant challenge in deep reinforcement learning (RL). To address this problem, researchers introduce the experience replay (ER) concept to complement the training of a deep RL agent. However, the buffer size, experience selection, and experience retention strategies adopted for the ER can negatively affect the agent’s performance stability, especially for complex continuous state action problems. This paper investigates how to address the stability problem using an enhanced ER… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 28 publications
(60 reference statements)
0
5
0
Order By: Relevance
“…However, its limited explainability, especially in Deep Reinforcement Learning (DRL), has hindered its adoption in regulated sectors like manufac-turing [2], finance [3], and health [4]. These challenges, including debugging, interpretation, and inefficiency, become more pronounced as Deep Learning advances, such as in improving Experience Replay [12][13][14][29][30][31][32]. While many have focused on improving Experience Replay's selection strategy, reducing its capacity remains largely unexplored.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…However, its limited explainability, especially in Deep Reinforcement Learning (DRL), has hindered its adoption in regulated sectors like manufac-turing [2], finance [3], and health [4]. These challenges, including debugging, interpretation, and inefficiency, become more pronounced as Deep Learning advances, such as in improving Experience Replay [12][13][14][29][30][31][32]. While many have focused on improving Experience Replay's selection strategy, reducing its capacity remains largely unexplored.…”
Section: Discussionmentioning
confidence: 99%
“…However, drawbacks included correlated samples, limited capacity causing an agent to forget information, outdated samples from non-stationary environments, and overfitting from samples memorized. Prioritized Experience Replay (PER) [29], Attention-Based Experience Replay [30], Combined Experience Replay (CER) [12], Explanation-Aware Experience Replay [31] and many more selection strategies [12][13][14]32] propose solutions to these drawbacks. Understanding Experience Replay is crucial for efficiency.…”
Section: Experience Replaymentioning
confidence: 99%
See 1 more Smart Citation
“…Wei et al [43], propose a quantum-inspired ER paradigm to achieve a better balance between exploration and exploitation. Recent works address the problem of catastrophic forgetting by proposing strategies to improve control over the mechanisms for storing, selecting, retaining, and forgetting in ER [44], including using transfer learning about past experiences [45]. Li et al [46] proposed the Self-generated Long-term Experience Replay (SLER) method to mitigate catastrophic forgetting in continuous learning tasks.…”
Section: Literature Review and Related Workmentioning
confidence: 99%
“…It is mainly related to buffer size limitation and is sensitive to experience sampling and storage strategies. Recent works addressed this problem by proposing strategies to improve control over the mechanisms for storing, selecting, retaining, and forgetting in experience replay (Osei and Lopez, 2023), including using transfer learning about past experiences (Anzaldo and Andrade, 2022). According to , their Self-generated Long-term Experience Replay (SLER) approach improved the dual experience replay algorithm, applied in continuous learning tasks to mitigate catastrophic forgetting by reducing its impact on computer memory-consuming growth.…”
Section: Research On Experience Replay and Some Directions For Future...mentioning
confidence: 99%