2018
DOI: 10.1007/978-3-319-98446-9_19
|View full text |Cite
|
Sign up to set email alerts
|

Building Collaboration in Multi-agent Systems Using Reinforcement Learning

Abstract: This paper presents a proof-of concept study for demonstrating the viability of building collaboration among multiple agents through standard Q learning algorithm embedded in particle swarm optimisation. Collaboration is formulated to be achieved among the agents via competition, where the agents are expected to balance their action in such a way that none of them drifts away of the team and none intervene any fellow neighbours territory, either. Particles are devised with Q learning for self training to learn… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 11 publications
(7 citation statements)
references
References 28 publications
0
7
0
Order By: Relevance
“…Developing collective behavior automatically for a group of robots is challenging (Francesca and Birattari, 2016 ). Although there are some existing works which simulate swarming collective behavior in robots, including collective navigation for robots (Na et al, 2022 ), collaborative robots (Aydin and Fellows, 2018 ), and collective formation of robots (Buffet et al, 2007 ), none of these can automatically generate a diverse set of collective behaviors. One limitation in doing this is that automatic recognition of swarming collective motion behavior is hard (Harvey et al, 2018 ).…”
Section: Background and Related Workmentioning
confidence: 99%
“…Developing collective behavior automatically for a group of robots is challenging (Francesca and Birattari, 2016 ). Although there are some existing works which simulate swarming collective behavior in robots, including collective navigation for robots (Na et al, 2022 ), collaborative robots (Aydin and Fellows, 2018 ), and collective formation of robots (Buffet et al, 2007 ), none of these can automatically generate a diverse set of collective behaviors. One limitation in doing this is that automatic recognition of swarming collective motion behavior is hard (Harvey et al, 2018 ).…”
Section: Background and Related Workmentioning
confidence: 99%
“…ToM-net [41] captures mental states of other agents and predicts their future action. OM [42] uses agent policies to predict the intended actions of opponents. But all these works are conducted under a competitive setting and require agents to infer each other's intention, which could be inaccurate considering the instability nature of MARL [43].…”
Section: Intention Modelingmentioning
confidence: 99%
“…Supervised learning is also used in works that incorporate Theory of Mind, Section II-D, which equips agents with a prediction module to estimate other agents' beliefs and future actions. In these cases, SL can be used to predict actions given current observations [38], [59], [94], [95] or coupled with the obverter technique to influence policy based on an agent's own understanding [96], [97].…”
Section: B: Gumbel Soft-max and Concrete Distributionmentioning
confidence: 99%
“…Alternatively, mental models can also be based on other agents' actions and perceptions without assuming similar belief systems. For instance, [94] augments agents' policy with predictions of other agents' behavior and demonstrates that agents can learn better policies using their estimates of other players' goals in cooperative and competitive situations. However, this work does not consider environments where communication is present.…”
Section: ) Modeling Agentsmentioning
confidence: 99%