2021
DOI: 10.48550/arxiv.2104.12895
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Computational Performance of Deep Reinforcement Learning to find Nash Equilibria

Abstract: We test the performance of deep deterministic policy gradient (DDPG)-a deep reinforcement learning algorithm, able to handle continuous state and action spaces-to learn Nash equilibria in a setting where firms compete in prices. These algorithms are typically considered "model-free" because they do not require transition probability functions (as in e.g., Markov games) or predefined functional forms. Despite being "model-free", a large set of parameters are utilized in various steps of the algorithm. These are… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 34 publications
0
1
0
Order By: Relevance
“…In particular, the team employed the deep deterministic policy gradient (DDPG) method [13,14]. Lanterne opted for DDPG since the team had already analyzed its suitability to find Nash equilbria in uniform price auctions and validated DDPG intensely against classical game theoretic predictions [15].…”
Section: Lanterne Rouge's Solutionmentioning
confidence: 99%
“…In particular, the team employed the deep deterministic policy gradient (DDPG) method [13,14]. Lanterne opted for DDPG since the team had already analyzed its suitability to find Nash equilbria in uniform price auctions and validated DDPG intensely against classical game theoretic predictions [15].…”
Section: Lanterne Rouge's Solutionmentioning
confidence: 99%