2020
DOI: 10.1016/j.neucom.2020.08.024
|View full text |Cite
|
Sign up to set email alerts
|

Random curiosity-driven exploration in deep reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 56 publications
(17 citation statements)
references
References 25 publications
0
17
0
Order By: Relevance
“…Directed exploration strategies utilize the previous history of the learning process and influence the portion of the environment explored in the future, including count-based [32], curiosity-driven [31,33], and upper confidence bounds (UCB) exploration [34]. For tabular-based RL, count-based exploration strategies give an extra exploration bonus to frequently visited states [32].…”
Section: B Related Workmentioning
confidence: 99%
“…Directed exploration strategies utilize the previous history of the learning process and influence the portion of the environment explored in the future, including count-based [32], curiosity-driven [31,33], and upper confidence bounds (UCB) exploration [34]. For tabular-based RL, count-based exploration strategies give an extra exploration bonus to frequently visited states [32].…”
Section: B Related Workmentioning
confidence: 99%
“…A growing number of researchers are applying DRL to solve the perception-decision problem in computer vision tasks [23]. [24] applies a policy gradient based on random curiosity-driven exploration to image coding, resulting in good performance in Atari games. Uzkent et al select HR image blocks with policy gradients to improve classification accuracy with LR images as states [25].…”
Section: B Computer Vision With Deep Reinforcement Learningmentioning
confidence: 99%
“…On the other hand, the disadvantage of the above path tracking methods is that all methods believe the robot model and external environment are ideal 21‐23 . Aiming at uncertain disturbances of complex robot systems, common methods include sliding mode control (SMC), active disturbance rejection control (ADRC), symplectic pseudospectral RHC controller, and adaptive control 24‐26 .…”
Section: Introductionmentioning
confidence: 99%