2019
DOI: 10.48550/arxiv.1903.03176
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MinAtar: An Atari-Inspired Testbed for Thorough and Reproducible Reinforcement Learning Experiments

Kenny Young,
Tian Tian

Abstract: The Arcade Learning Environment (ALE) is a popular platform for evaluating reinforcement learning agents. Much of the appeal comes from the fact that Atari games demonstrate aspects of competency we expect from an intelligent agent and are not biased toward any particular solution approach. The challenge of the ALE includes (1) the representation learning problem of extracting pertinent information from raw pixels, and (2) the behavioural learning problem of leveraging complex, delayed associations between act… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
58
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 25 publications
(58 citation statements)
references
References 8 publications
0
58
0
Order By: Relevance
“…We select the MinAtar (Young and Tian, 2019) environments to test the proposed Deep RL algorithms. MinAtar is a testbed incorporating several Atari games from the Arcade Learning Environment (Bellemare et al, 2013) with a reduced state-representation.…”
Section: Minatarmentioning
confidence: 99%
See 3 more Smart Citations
“…We select the MinAtar (Young and Tian, 2019) environments to test the proposed Deep RL algorithms. MinAtar is a testbed incorporating several Atari games from the Arcade Learning Environment (Bellemare et al, 2013) with a reduced state-representation.…”
Section: Minatarmentioning
confidence: 99%
“…The platform incorporates sticky actions (Machado et al, 2018) and is designed to enable thorough algorithmic comparisons due to reduced computation times. Following Young and Tian (2019), the network structure consists of a convolutional and a fully-connected layer. The remaining hyperparameters match Young and Tian (2019), except that we use the Adam (Kingma and Ba, 2014) optimizer, which led to much more stable results during our experiments.…”
Section: Minatarmentioning
confidence: 99%
See 2 more Smart Citations
“…Authors of [12] explored this phenomenon in detail for both DQN and PPO algorithm. However, they studied it on small MinAtar subtasks [13] (MinAtar is a simplified version of classic AtariGames).…”
Section: Pruning Lottery Ticketmentioning
confidence: 99%