2017
DOI: 10.48550/arxiv.1709.06009
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Revisiting the Arcade Learning Environment: Evaluation Protocols and Open Problems for General Agents

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
28
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 15 publications
(28 citation statements)
references
References 0 publications
0
28
0
Order By: Relevance
“…We empirically study the behavior of BDQN on a wide range of Atari games [16,44]. Since BDQN follows an efficient exploration-exploitation strategy, it reaches much higher cumulative rewards in fewer interactions, compared to its ε-greedy predecessor DDQN.…”
Section: Strategymentioning
confidence: 99%
“…We empirically study the behavior of BDQN on a wide range of Atari games [16,44]. Since BDQN follows an efficient exploration-exploitation strategy, it reaches much higher cumulative rewards in fewer interactions, compared to its ε-greedy predecessor DDQN.…”
Section: Strategymentioning
confidence: 99%
“…Ruder (2017) provides a survey of multi-task learning in general, which, different from our problem of interest, considers a fixed finite population of tasks. Finn et al (2017) (Brockman et al, 2016) (which we build upon), the Arcade Learning Environment (Bellemare et al, 2013;Machado et al, 2017), DeepMind Lab (Beattie et al, 2016), and VizDoom (Kempka et al, 2016). The MuJoCo physics simulator (Todorov et al, 2012) has been influential in standardizing a number of continuous control tasks.…”
Section: Related Workmentioning
confidence: 99%
“…The low computation cost of Pong allows us to do an extensive study of the bias-variance of Q for different model-based planning and exploration strategies. (Even for this game, GATS requires multiple weeks of GPU process for a short run of 5M time steps) Experimental results For this work, we also developed a new OpenAI gym (Brockman et al, 2016)-like interface for the latest Atari Learning Environment (ALE) (Machado et al, 2017), which supports different modes and difficulties for Atari games. We study the sample complexity required by GDM and RP to adapt and transfer from one domain of the game (a mode and difficulty) to another domain (another mode and difficulty).…”
Section: Technical Contributionsmentioning
confidence: 99%