2015
DOI: 10.1007/s10462-015-9447-5
|View full text |Cite
|
Sign up to set email alerts
|

Exponential moving average based multiagent reinforcement learning algorithms

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2018
2018
2025
2025

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(13 citation statements)
references
References 29 publications
0
13
0
Order By: Relevance
“…In this work, the exponential moving average (EMA) is adopted as a low-pass filter. The EMA method is model-free and has been widely used in time series analyses [ 30 ]. In EMA, recent data points have higher weights than older ones [ 31 ].…”
Section: Methodsmentioning
confidence: 99%
“…In this work, the exponential moving average (EMA) is adopted as a low-pass filter. The EMA method is model-free and has been widely used in time series analyses [ 30 ]. In EMA, recent data points have higher weights than older ones [ 31 ].…”
Section: Methodsmentioning
confidence: 99%
“…In the experiment, EMA (exponential moving average) Q-learning [11], WoLF-PHC [10], and SARSA (stateaction-reward-state-action) [21] are chosen as comparison algorithms. EMA Q-learning and WoLF-PHC are MARL algorithms while SARSA is a type of single-agent RL algorithm corresponding to centralized learning in the context of multiple agents.…”
Section: Simulations On Stochastic Gamesmentioning
confidence: 99%
“…WoLF-policy hillclimbing (WoLF-PHC) [10] only needed to share states and local immediate rewards of each agent, but the convergence property was not guaranteed any more. The exponential moving average (EMA) Q-learning [11] and the weighted policy learner (WPL) [12] empirically converged to a Nash equilibrium in some typical repeated games. To design scalable MARL algorithms that can gain the optimal total sum of reward in fully cooperative games is our motivation.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In this chapter, we propose two MARL algorithms. The algorithms proposed in this chapter have already been published in [2][3][4]. The first proposed algorithm can successfully converge to Nash equilibrium policies in games that have pure Nash equilibrium.…”
Section: Introductionmentioning
confidence: 99%