2021
DOI: 10.1007/s40747-021-00345-6
|View full text |Cite
|
Sign up to set email alerts
|

A game strategy model in the digital curling system based on NFSP

Abstract: The digital curling game is a two-player zero-sum extensive game in a continuous action space. There are some challenging problems that are still not solved well, such as the uncertainty of strategy, the large game tree searching, and the use of large amounts of supervised data, etc. In this work, we combine NFSP and KR-UCT for digital curling games, where NFSP uses two adversary learning networks and can automatically produce supervised data, and KR-UCT can be used for large game tree searching in continuous … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(12 citation statements)
references
References 9 publications
0
12
0
Order By: Relevance
“…As we can see, we have a higher winning rate against the other two models because our Curling MCTS can explore new action spaces via separate alternative options. It is worth mentioning that our decisionmaking method has a lower winning rate when compared to NFSP [21]. This is because NFSP uses two adversary learning networks to produce supervised data automatically.…”
Section: Resultsmentioning
confidence: 97%
See 2 more Smart Citations
“…As we can see, we have a higher winning rate against the other two models because our Curling MCTS can explore new action spaces via separate alternative options. It is worth mentioning that our decisionmaking method has a lower winning rate when compared to NFSP [21]. This is because NFSP uses two adversary learning networks to produce supervised data automatically.…”
Section: Resultsmentioning
confidence: 97%
“…The KR-DL-UCT searched in continuous action space by kernel regression and trained by supervised learning and self-game reinforcement learning, which improved the speed and accuracy of constant Monte Carlo tree search and won the international digital curling competition. Han et al [21] combined NFSP and KR-UCT in a digital curling game. The NFSP uses two adversary learning networks to produce supervised data, and KR-UCT can be used for large game tree searching in continuous action space.…”
Section: Digital Curling Policy Recommendationmentioning
confidence: 99%
See 1 more Smart Citation
“…Currently, deep reinforcement learning (DRL) has shown tremendous potential in solving complex decision-making problems. DRL has spurred a lot of significant breakthroughs in many applications, such as manipulator controlling [18], autonomous driving [19,20], and games [21,22]. Considering the strengths of DRL, researchers attempt to apply it to tackle the PointGoal navigation problem [4,[23][24][25][26].…”
Section: Introductionmentioning
confidence: 99%
“…The RL approach has been used to find out the game strategy in digital curling by reaching the Nash equilibrium [8]. The digital curling game is a two-player zero-sum extensive game in a continuous action space.…”
Section: Introductionmentioning
confidence: 99%