2014
DOI: 10.1007/978-3-319-13987-6_34
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Stage Temporal Difference Learning for 2048

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 11 publications
0
7
0
Order By: Relevance
“…Note that this technique is different from the one introduced for 2048 by Wu et al [38]. Their algorithm learns the n-tuple networks stage by stage.…”
Section: E Carousel Shaping and Redundant Encodingmentioning
confidence: 95%
See 3 more Smart Citations
“…Note that this technique is different from the one introduced for 2048 by Wu et al [38]. Their algorithm learns the n-tuple networks stage by stage.…”
Section: E Carousel Shaping and Redundant Encodingmentioning
confidence: 95%
“…Their best function involved nearly 23 million parameters and scored 100 178 on average at 1-ply. Wu et al [38] has later extended this work by employing TD(0) with 5 million training games to learn a larger systematic n-tuple system with 67 million parameters, scoring 142 727 on average. By employing three such ntuple networks enabled at different stages of the game along with expectimax with depth 5, they were able to achieve 328 946 points on average.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Note that the preliminary version of this paper [32] did not include the following: MS-TD applied to Threes, more splitting strategies for both 2048 and Threes, and a demonstration of combining MS-TD with other techniques to improve 2048 and Threes programs.…”
Section: Introductionmentioning
confidence: 99%