2022
DOI: 10.48550/arxiv.2202.05841
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Entropic fictitious play for mean field optimization problem

Zhenjie Ren,
Songbo Wang

Abstract: It is well known that the training of the neural network can be viewed as a mean field optimization problem. In this paper we are inspired by the fictitious play, a classical algorithm in the game theory for learning the Nash equilibria, and propose a new algorithm, different from the conventional gradient-descent ones, to solve the mean field optimization. We rigorously prove its (exponential) convergence, and show some simple numerical examples.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 8 publications
0
4
0
Order By: Relevance
“…Remark 1. Under Assumption 1, the existence and uniqueness of the minimizer µ * ∈ P 2 of L is guaranteed by Ren and Wang (2022), and µ * satisfies the first-order optimality condition: µ * = μ * (Hu et al, 2019).…”
Section: Problem Setupmentioning
confidence: 99%
See 2 more Smart Citations
“…Remark 1. Under Assumption 1, the existence and uniqueness of the minimizer µ * ∈ P 2 of L is guaranteed by Ren and Wang (2022), and µ * satisfies the first-order optimality condition: µ * = μ * (Hu et al, 2019).…”
Section: Problem Setupmentioning
confidence: 99%
“…The entropic fictitious play (EFP) algorithm (Ren and Wang, 2022) is the optimization method that minimizes the objective (4). The continuous evolution {µ t } t≥0 ⊂ P 2 of the EFP is defined as follows: for γ > 0,…”
Section: The Ideal Updatementioning
confidence: 99%
See 1 more Smart Citation
“…Such optimization problems have attracted considerable attention in recent years, see e.g. [9,16,12,18,7]. In this setting, there exist multiple different choices of flows of probability measures (m t ) t≥0 that can serve as analogues of the gradient descent algorithm in R d , as well as multiple different choices of conditions on V analogous to (1.1) that can be used to prove convergence of such flows.…”
Section: Introductionmentioning
confidence: 99%