2013
DOI: 10.1613/jair.3904
|View full text |Cite
|
Sign up to set email alerts
|

Decentralized Anti-coordination Through Multi-agent Learning

Abstract: To achieve an optimal outcome in many situations, agents need to choose distinct actions from one another. This is the case notably in many resource allocation problems, where a single resource can only be used by one agent at a time. How shall a designer of a multi-agent system program its identical agents to behave each in a different way? From a game theoretic perspective, such situations lead to undesirable Nash equilibria. For example consider a resource allocation game in that two players compete for an … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(25 citation statements)
references
References 24 publications
0
25
0
Order By: Relevance
“…These available actions have an associated payoff, which rates how beneficial or detrimental the "movement" for each player is. The game itself only describes what the players can do, but not the ultimate actions, in the same way the model equations constrain the variables feasible values [36], [83], [84]. Every strategy followed by one player generates a benefit for the named agent and a loss for the rest, which is the socalled payoff.…”
Section: F Game Theorymentioning
confidence: 99%
“…These available actions have an associated payoff, which rates how beneficial or detrimental the "movement" for each player is. The game itself only describes what the players can do, but not the ultimate actions, in the same way the model equations constrain the variables feasible values [36], [83], [84]. Every strategy followed by one player generates a benefit for the named agent and a loss for the rest, which is the socalled payoff.…”
Section: F Game Theorymentioning
confidence: 99%
“…However, there are certain assumptions these papers have in common that hinder the accurate evaluation of ML solutions. For instance, most papers consider synchronous time slots (e.g., [20]- [25]), which is a strong assumption that does not hold at all in carrier sense multiple access with collision avoidance (CSMA/CA) WLANs. Also, some papers define a binary reward, where actions are simply good or bad (e.g., [20], [24], [26]).…”
Section: Related Workmentioning
confidence: 99%
“…For instance, most papers consider synchronous time slots (e.g., [20]- [25]), which is a strong assumption that does not hold at all in carrier sense multiple access with collision avoidance (CSMA/CA) WLANs. Also, some papers define a binary reward, where actions are simply good or bad (e.g., [20], [24], [26]). This, while easing the RL framework, hinders rewards taking into account continuous-valued performance metrics like throughput or delay.…”
Section: Related Workmentioning
confidence: 99%
“…There exist no research work done until yet for smart decision making for fusion center, currently AND/OR logic is available to take decision at fusion center.in the next section of paper we will show experimental results comparing our approach with traditional AND/OR logic approach. After going through a lot of research paper Random Forest seems to be most suitable in deployment of 5G Network [61]-[64], reason is accuracy and capability to become classifier of future generation. Since, 5G is also futuristic technology it require a strong and more accurate classifier.…”
Section: B Working Of Random Forest Algorithmmentioning
confidence: 99%