Proceedings of the 25th International Conference on Machine Learning - ICML '08 2008
DOI: 10.1145/1390156.1390202
|View full text |Cite
|
Sign up to set email alerts
|

No-regret learning in convex games

Abstract: Quite a bit is known about minimizing different kinds of regret in experts problems, and how these regret types relate to types of equilibria in the multiagent setting of repeated matrix games. Much less is known about the possible kinds of regret in online convex programming problems (OCPs), or about equilibria in the analogous multiagent setting of repeated convex games. This gap is unfortunate, since convex games are much more expressive than matrix games, and since many important machine learning problems … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
66
0

Year Published

2008
2008
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 49 publications
(67 citation statements)
references
References 9 publications
1
66
0
Order By: Relevance
“…While there are many algorithms that are guaranteed to achieve no regret (e.g., Bowling, 2004;Foster & Vohra, 1999;Gordon et al, 2008), we show that it is impossible for an algorithm to be guaranteed to have no disappointment against an unknown associate. However, it is possible for an algorithm to quickly achieve and maintain low disappointment against classes of algorithms in many repeated games.…”
Section: Introductionmentioning
confidence: 81%
See 2 more Smart Citations
“…While there are many algorithms that are guaranteed to achieve no regret (e.g., Bowling, 2004;Foster & Vohra, 1999;Gordon et al, 2008), we show that it is impossible for an algorithm to be guaranteed to have no disappointment against an unknown associate. However, it is possible for an algorithm to quickly achieve and maintain low disappointment against classes of algorithms in many repeated games.…”
Section: Introductionmentioning
confidence: 81%
“…Agent i is said to have no regret when lim T →∞R T i ≤ 0. When all agents use no-regret learning rules, play converges to correlated equilibria (Greenwald & Jafari, 2003;Gordon et al, 2008).…”
Section: Regretmentioning
confidence: 99%
See 1 more Smart Citation
“…Other equilibrium concepts that provide computational challenges include correlated equilibria, cooperative equilibria, and evolutionary stable strategies [14,145,423]. Other topics of interest are learning in repeated games and regret minimizing algorithms [146,153,174] and games with state and their connections to reinforcement learning [50,61,229]. Finally, cryptographic methods are needed to make sure that in the course of a game, a player's views, utilities, preferences, strategies are kept private.…”
Section: Algorithmic Issues In Game Theory and Computer Science 19mentioning
confidence: 99%
“…In fact, this is an impossible task if one does not add priors, which is equivalent to adding structure on the environment. Since Nature's complexity is unbounded, even a very patient 1 Littlestone and Warmuth (1994); Cesa-Bianchi et al (1996); Vovk (1998); Auer and Long (1999); Foster and Vohra (1999); Freund and Schapire (1999); Lugosi (2003, 2006); Greenwald and Jafari (2003); Cesa-Bianchi et al (2007); Gordon et al (2008). 2 Hannan (1957); Foster and Vohra (1993, 1997, 1998; Levine (1995, 1999);Hart andMas-Colell (2000, 2001a); Lehrer (2003); Hart (2005).…”
mentioning
confidence: 99%