1996
DOI: 10.1016/0004-3702(95)00096-8
|View full text |Cite
|
Sign up to set email alerts
|

Best-first minimax search

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

1997
1997
2023
2023

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 44 publications
(5 citation statements)
references
References 20 publications
0
5
0
Order By: Relevance
“…We now present the framework of the Descent algorithm (Cohen-Solal, 2020). The learning framework of Descent is based on Unbounded Minimax (Korf and Chickering, 1996): an algorithm calculating an approximation of the minimax value of a game state ; and on Descent Minimax: a variant of Unbounded Minimax which consists in exploring the sequences of actions until terminal states. In comparison, Unbounded Minimax and MCTS explore a sequence of actions only until a leaf state is reached.…”
Section: Descent Minimaxmentioning
confidence: 99%
“…We now present the framework of the Descent algorithm (Cohen-Solal, 2020). The learning framework of Descent is based on Unbounded Minimax (Korf and Chickering, 1996): an algorithm calculating an approximation of the minimax value of a game state ; and on Descent Minimax: a variant of Unbounded Minimax which consists in exploring the sequences of actions until terminal states. In comparison, Unbounded Minimax and MCTS explore a sequence of actions only until a leaf state is reached.…”
Section: Descent Minimaxmentioning
confidence: 99%
“…PNS explores in first the node with the lowest proof number. Best-First Search [Korf and Chickering, 1994] calls the evaluation function for all child nodes and explores the best node first. SSS* [Stockman, 1979] explores all nodes in parallel as A* would do it with a specific heuristic.…”
Section: Best First Searchmentioning
confidence: 99%
“…Unlike AlphaZero-like algorithms (Silver et al, 2018), the Descent framework uses a variant of Unbounded Minimax (Korf and Chickering, 1996), instead of Monte Carlo Tree Search, to construct the partial game tree used to determine the best action to play and to collect data for learning. During training, at each move, the best sequences of moves are iteratively extended until terminal states.…”
mentioning
confidence: 99%