2014
DOI: 10.1080/09540091.2014.885294
|View full text |Cite
|
Sign up to set email alerts
|

A framework for learning and planning against switching strategies in repeated games

Abstract: Intelligent agents, human or artificial, often change their behaviour as they interact with other agents. For an agent to optimise its performance when interacting with such agents, it must be capable of detecting and adapting according to such changes. This work presents an approach on how to effectively deal with non-stationary switching opponents in a repeated game context. Our main contribution is a framework for online learning and planning against opponents that switch strategies. We present how two oppo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2014
2014
2017
2017

Publication Types

Select...
3
2

Relationship

3
2

Authors

Journals

citations
Cited by 13 publications
(15 citation statements)
references
References 16 publications
0
15
0
Order By: Relevance
“…4.5). Comparisons are performed with state of the art approaches: two of our previous approaches MDP4.5 [25] and MDP-CL [26]; R-max [7] used as baseline; FAL [18] since it is a fast learning algorithm in repeated games, WOLF-PHC [6] 3 since it can learn non-stationary environments; and the omniscient (perfect) agent that best responds immediately to switches. Results are compared in terms of average utility over the repeated game.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…4.5). Comparisons are performed with state of the art approaches: two of our previous approaches MDP4.5 [25] and MDP-CL [26]; R-max [7] used as baseline; FAL [18] since it is a fast learning algorithm in repeated games, WOLF-PHC [6] 3 since it can learn non-stationary environments; and the omniscient (perfect) agent that best responds immediately to switches. Results are compared in terms of average utility over the repeated game.…”
Section: Methodsmentioning
confidence: 99%
“…Another approach that uses MDPs to represent opponent strategies is the MDP-CL approach [26]. We introduced MDP-CL in previous work to act against non-stationary opponents (see Algorithm 1).…”
Section: Algorithm 1: Mdp-cl [26]mentioning
confidence: 99%
See 3 more Smart Citations