2008
DOI: 10.1007/978-3-540-85565-1_53
|View full text |Cite
|
Sign up to set email alerts
|

Learning Grouping and Anti-predator Behaviors for Multi-agent Systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2010
2010
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(10 citation statements)
references
References 9 publications
0
10
0
Order By: Relevance
“…DEVS capabilities in the field of two practical examples, namely, a model of unmanned aerial vehicles and a model of an automated systemare presented.The described ________________________________________________________________________________________________ 26 defence models have been previously presented and well-received at the Baltic Defence Research and Technology Conference, Riga,September [10][11]2009. Although additional experiments might be further required, computational results conducted so far clearly show that the V-DEVS approach is suitable for the development and analysis of defence models.Future along this line of research includes improvement and validation of the described simulation algorithms and models.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…DEVS capabilities in the field of two practical examples, namely, a model of unmanned aerial vehicles and a model of an automated systemare presented.The described ________________________________________________________________________________________________ 26 defence models have been previously presented and well-received at the Baltic Defence Research and Technology Conference, Riga,September [10][11]2009. Although additional experiments might be further required, computational results conducted so far clearly show that the V-DEVS approach is suitable for the development and analysis of defence models.Future along this line of research includes improvement and validation of the described simulation algorithms and models.…”
Section: Discussionmentioning
confidence: 99%
“…The reinforcement learningapproach, namely Q-Learning,is used for the search of a nearly optimal solution in the patrolling area. The reinforcement learning is a machine learning approach in which transition rules are learned through the experience of agents in their environment.Q-learning [10]is known as the best understood reinforcement learning technique used for maximizing the sum of the rewards received.In Q-Learning, the learning process consists of acquiring a state t s , deciding an action t a , receiving a reward r from an environment, and updating Q-value ( ) , t t Q s a [11] calculated by the following equation:…”
Section: A Model Of Unmanned Aerial Vehiclesmentioning
confidence: 99%
“…They find that the swarming among prey is influenced by how predators attack. While Reynolds (1999) proposes static rules leading to swarming, Morihiro et al (2008) have their agents learn to do so by encoding these rules in the reward of an RL algorithm. Further, Sunehag et al (2019) find the emergence of flocking and symbiosis with rewards shaping by independent MARL in simulated multiple-species ecosystems.…”
Section: Further Emergent Behaviormentioning
confidence: 99%
“…Supplementary forces can be introduced, which repel an individual from an enemy or from obstacles, for example. To overcome these static rules definitions, Morihiro et al (2008) used Reinforcement Learning, particularly Q-Learning, to train agents to follow these rules. In their model the agents iteratively learn while at every time step an agent i only considers one other agent j.…”
Section: Swarm Behaviormentioning
confidence: 99%
“…The question arises whether flocking behavior occurs in a scenario with such properties where agents solely try to maximize their survival time. In contrast to Morihiro et al (2008), we pursue a scenario in which agents are trained with reinforcement learning solely on the objective to survive, without explicitly enforcing swarm behavior. Additionally, we demonstrate that SELFish also works for a continuous action space of the agents.…”
Section: Swarm Behaviormentioning
confidence: 99%