2011
DOI: 10.1007/s13235-011-0038-z
|View full text |Cite
|
Sign up to set email alerts
|

Distributed Dynamic Reinforcement of Efficient Outcomes in Multiagent Coordination and Network Formation

Abstract: We analyze reinforcement learning under so-called "dynamic reinforcement." In reinforcement learning, each agent repeatedly interacts with an unknown environment (i.e., other agents), receives a reward, and updates the probabilities of its next action based on its own previous actions and received rewards. Unlike standard reinforcement learning, dynamic reinforcement uses a combination of long-term rewards and recent rewards to construct myopically forward looking action selection probabilities. We analyze the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
49
0

Year Published

2012
2012
2017
2017

Publication Types

Select...
5
3
1

Relationship

3
6

Authors

Journals

citations
Cited by 30 publications
(50 citation statements)
references
References 41 publications
1
49
0
Order By: Relevance
“…Game theory has important applications to the design and control of multiagent systems [1]- [9]. This design choice requires two steps.…”
Section: Introductionmentioning
confidence: 99%
“…Game theory has important applications to the design and control of multiagent systems [1]- [9]. This design choice requires two steps.…”
Section: Introductionmentioning
confidence: 99%
“…To this end, we employ a learning framework (namely, perturbed learning automata) that is based on the reinforcement learning algorithm introduced by the authors in [6,7]. It belongs to the general class of learning automata [13].…”
Section: Reinforcement Learning (Rl)mentioning
confidence: 99%
“…The concept is inspired by classical methods in feedback control as well as the psychological tendency to extrapolate from past trends. Anticipatory learning was utilized in [3,20,21], where it was shown how anticipatory learning can alter the convergence to both mixed and pure equilibria.…”
Section: Smoothed Payoff Modificationmentioning
confidence: 99%