2016 ACM/IEEE 7th International Conference on Cyber-Physical Systems (ICCPS) 2016
DOI: 10.1109/iccps.2016.7479108
|View full text |Cite
|
Sign up to set email alerts
|

On Learning How Players Learn: Estimation of Learning Dynamics in the Routing Game

Abstract: The routing game models congestion in transportation networks, communication networks, and other cyber physical systems in which agents compete for shared resources. We consider an online learning model of player dynamics: at each iteration, every player chooses a route (or a probability distribution over routes, which corresponds to a flow allocation over the physical network), then the joint decision of all players determines the costs of each path, which are then revealed to the players.We pose the followin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
4
2

Relationship

1
9

Authors

Journals

citations
Cited by 19 publications
(10 citation statements)
references
References 27 publications
0
10
0
Order By: Relevance
“…To account for this, a growing corpus of literature has been devoted to studying the problem of equilibrium convergence under no-regret learning, typically focusing on special classes of games (such as convex potential games, zero-sum games, routing games, etc.). Here again, most efforts have focused on the convergence of the time-averaged sequence of play because of its connection to the players' regret (Krichene et al 2015, Balandat et al 2016, Lam et al 2016). Nevertheless, the convergence of the actual sequence of play is also crucial to investigate for several reasons: First, from a practical standpoint, since the players' payoffs are determined by the sequence of chosen actions and not some fictitious variant thereof, convergence to Nash equilibrium should also be stated in terms of the actual sequence of play.…”
Section: Our Contributions and Related Workmentioning
confidence: 99%
“…To account for this, a growing corpus of literature has been devoted to studying the problem of equilibrium convergence under no-regret learning, typically focusing on special classes of games (such as convex potential games, zero-sum games, routing games, etc.). Here again, most efforts have focused on the convergence of the time-averaged sequence of play because of its connection to the players' regret (Krichene et al 2015, Balandat et al 2016, Lam et al 2016). Nevertheless, the convergence of the actual sequence of play is also crucial to investigate for several reasons: First, from a practical standpoint, since the players' payoffs are determined by the sequence of chosen actions and not some fictitious variant thereof, convergence to Nash equilibrium should also be stated in terms of the actual sequence of play.…”
Section: Our Contributions and Related Workmentioning
confidence: 99%
“…This point was only recently expounded rigorously, by Mertikopoulos et al (2018b), who show that even though follow-the-regularizedleader (another no-regret learning algorithm) converges to a Nash equilibrium in linear zero-sum games in the sense of time averages, actual joint actions orbit Nash equilibria in perpetuity. Motivated by this consideration, a growing literature (Krichene et al, 2015;Lam et al, 2016;Palaiopanos et al, 2017;Zhou et al, 2017;Mertikopoulos et al, 2017;Zhou et al, 2018;Mertikopoulos & Zhou, 2019) has aimed at obtaining last-iterate convergence results. However, all of these last-iterate convergence results are qualitative.…”
Section: Related Workmentioning
confidence: 99%
“…In [21], the routing game, based on non-cooperative game theory, has been explored for multiple agents to learn about the behaviour of other agents in order to make routing decisions about the use of limited shared resources (e.g., roads).…”
Section: A Vehicle-to-vehicle Cooperation and Reasoningmentioning
confidence: 99%