2001
DOI: 10.1007/3-540-44795-4_33
|View full text |Cite
|
Sign up to set email alerts
|

Social Agents Playing a Periodical Policy

Abstract: Coordination is an important issue in multiagent systems. Within the stochastic game framework this problem translates to policy learning in a joint action space. This technique however suffers some important drawbacks like the assumption of the existence of a unique Nash equilibrium and synchronicity, the need for central control, the cost of communication, etc. Moreover in general sum games it is not always clear which policies should be learned. Playing pure Nash equilibria is often unfair to at least one o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2002
2002
2010
2010

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(8 citation statements)
references
References 5 publications
0
8
0
Order By: Relevance
“…In a common interest game, ESRL is able to find one of the Pareto optimal solutions of the game. In a conflicting interest game, we show that ESRL agents learn optimal fair, possibly periodical policies [17,26]. Important to know is that ESRL agents are independent in the sense that they only use their own action choices and rewards to base their decisions on, that ESRL agents are flexible in learning different solution concepts and they can handle both stochastic, possible delayed rewards and asynchronous action selection.…”
Section: Introductionmentioning
confidence: 98%
See 1 more Smart Citation
“…In a common interest game, ESRL is able to find one of the Pareto optimal solutions of the game. In a conflicting interest game, we show that ESRL agents learn optimal fair, possibly periodical policies [17,26]. Important to know is that ESRL agents are independent in the sense that they only use their own action choices and rewards to base their decisions on, that ESRL agents are flexible in learning different solution concepts and they can handle both stochastic, possible delayed rewards and asynchronous action selection.…”
Section: Introductionmentioning
confidence: 98%
“…We call a solution optimally fair when there is no other solution that is also fair for the agents but gives the agents more reward on average. Periodical policies were first introduced in [17].…”
Section: Introductionmentioning
confidence: 99%
“…We compare our results with a related periodical policy [5], and simulations show that agents using our adaptive strategy are able to achieve more optimal fairness results in the sense of obtaining higher utilitarian social welfare. Besides, the agents using our adaptive strategy can achieve fairness with less payoff cost compared with periodical policy when period length becomes smaller.…”
Section: Conclusion and Further Workmentioning
confidence: 96%
“…Nowé et al [5] [7] propose a periodical policy for achieving fair outcomes among multiple agents in a distributed way. This policy can be divided into two periods: reinforcement learning period and communication period.…”
Section: B Fairness Through Multi-agent Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation