ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021
DOI: 10.1109/icassp39728.2021.9414476
|View full text |Cite
|
Sign up to set email alerts
|

Solving a Class of Non-Convex Min-Max Games Using Adaptive Momentum Methods

Abstract: Adaptive momentum methods have recently attracted a lot of attention for training of deep neural networks. They use an exponential moving average of past gradients of the objective function to update both search directions and learning rates. However, these methods are not suited for solving min-max optimization problems that arise in training generative adversarial networks. In this paper, we propose an adaptive momentum min-max algorithm that generalizes adaptive momentum methods to the non-convex min-max re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0
2

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 12 publications
(15 citation statements)
references
References 13 publications
0
13
0
2
Order By: Relevance
“…• Non-monotone case: In this setting the convergence is guaranteed up to some accuracy that is governed by the stochastic nature of the problem (the σ 2 term) and by the distributed nature of the problem (∆ terms). With this respect the results are similar to non-distributed stochastic extragradient method [3] and distributed method in the homogeneous case [37]. To the best of our knowledge, convergence up to arbitrarily small accuracy can be guaranteed only for deterministic distributed methods [38], i.e.…”
Section: Resultsmentioning
confidence: 67%
“…• Non-monotone case: In this setting the convergence is guaranteed up to some accuracy that is governed by the stochastic nature of the problem (the σ 2 term) and by the distributed nature of the problem (∆ terms). With this respect the results are similar to non-distributed stochastic extragradient method [3] and distributed method in the homogeneous case [37]. To the best of our knowledge, convergence up to arbitrarily small accuracy can be guaranteed only for deterministic distributed methods [38], i.e.…”
Section: Resultsmentioning
confidence: 67%
“…The focus of this paper is on the decentralized problem (Algorithm 2) and we only mention the most related corollary for Algorithm 1 in the sequel. Theoretical results and performance of Adam 3 on a single computing node are given in [85].…”
Section: A Decentralized Adaptive Momentum Algorithmmentioning
confidence: 99%
“…Corollary 4.1. (Rephrased from [85]) Algorithm 1 requires a total of O( −4 ) gradient evaluations of the objective function to find an -SFNE point of Game 1.…”
Section: A Decentralized Adaptive Momentum Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…As such, existing positive results rely on adding this structure in one way or another. In particular, a common methodology is to "mimic" the case of a VI with a monotone operator [28] by imposing pseudomonotonicity, the Minty condition, or contraction of the best response mappings [67,68,69,70,71,72,73,74,75,76]. However, these assumptions are restrictive and rarely satisfied in modern applications; see [48] for a detailed discussion.…”
Section: Introductionmentioning
confidence: 99%