2023
DOI: 10.1007/978-3-031-25599-1_18
|View full text |Cite
|
Sign up to set email alerts
|

MicroRacer: A Didactic Environment for Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 17 publications
0
2
0
Order By: Relevance
“…In contrast to discretization methods for managing motion actions, DRL approaches like the DDPG learn policies directly without explicit discretization. However, the DDPG poses challenges in validation and reproducibility for continuous actions, varying implementations leading to inconsistent performance and generalization [34], catastrophic forgetting resulting in large fluctuations in training [35]. A3C [34], employing asynchronous gradient descent with concurrent environment instances, enhances training stability through parallel agents, while A2C [34], its synchronous counterpart, balances exploration-exploitation with a focus on uncertain value estimate states, often measured using entropy, and both demonstrate comparable performance in various scenarios.…”
Section: State Actor-critic Methodsmentioning
confidence: 99%
“…In contrast to discretization methods for managing motion actions, DRL approaches like the DDPG learn policies directly without explicit discretization. However, the DDPG poses challenges in validation and reproducibility for continuous actions, varying implementations leading to inconsistent performance and generalization [34], catastrophic forgetting resulting in large fluctuations in training [35]. A3C [34], employing asynchronous gradient descent with concurrent environment instances, enhances training stability through parallel agents, while A2C [34], its synchronous counterpart, balances exploration-exploitation with a focus on uncertain value estimate states, often measured using entropy, and both demonstrate comparable performance in various scenarios.…”
Section: State Actor-critic Methodsmentioning
confidence: 99%
“…Diffusion probabilistic models are outperforming existing state-of-the-art methods in various fields and applications [10,11]. Recent advancements in image-to-image translation tasks, such as inpainting and colorization, have showcased the potential of diffusion models to achieve superior performance compared to other state-of-the-art generative architectures.…”
Section: State Of the Artmentioning
confidence: 99%