2020
DOI: 10.1103/physreve.102.012601
|View full text |Cite
|
Sign up to set email alerts
|

Learning to flock through reinforcement

Abstract: Flocks of birds, schools of fish, insects swarms are examples of coordinated motion of a group that arises spontaneously from the action of many individuals. Here, we study flocking behavior from the viewpoint of multi-agent reinforcement learning. In this setting, a learning agent tries to keep contact with the group using as sensory input the velocity of its neighbors. This goal is pursued by each learning individual by exerting a limited control on its own direction of motion. By means of standard reinforce… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
44
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 43 publications
(46 citation statements)
references
References 27 publications
2
44
0
Order By: Relevance
“…[53,63]) consider a variant of the above distribution that only follows the PL form for steps longer than some threshold ℓ 0 , for example when analysing experimental data that become increasingly noisy at short step-lengths. However, since the step lengths resulting from our simulations are natively discrete, the unbounded PL distribution given in Eq (14) seems appropriate. Moreover, if one were to introduce a lower bound ℓ 0 > 1, one would need to add more parameters in the model to account for the probabilities p(ℓ) for all 1 � ℓ < ℓ 0 , which we consider an unnecessary complication.…”
Section: Plos Onementioning
confidence: 99%
See 2 more Smart Citations
“…[53,63]) consider a variant of the above distribution that only follows the PL form for steps longer than some threshold ℓ 0 , for example when analysing experimental data that become increasingly noisy at short step-lengths. However, since the step lengths resulting from our simulations are natively discrete, the unbounded PL distribution given in Eq (14) seems appropriate. Moreover, if one were to introduce a lower bound ℓ 0 > 1, one would need to add more parameters in the model to account for the probabilities p(ℓ) for all 1 � ℓ < ℓ 0 , which we consider an unnecessary complication.…”
Section: Plos Onementioning
confidence: 99%
“…This type of agent-based models that employ artificial intelligence to model behavior are gaining popularity in the last few years. Artificial neural networks (ANN) have been used, for instance, in the context of navigation behaviors [10,11] and reinforcement learning (RL) algorithms have been applied to model collective behavior in different scenarios, such as pedestrian movement [12] or flocking [13,14].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…As stated above, LEUP circumvents the biophysical details of cell migration. The need to model systems of interacting agents without previous knowledge of the biophysical mechanisms involved has sparked at least another agent based model 41 . In this model, similarly to ours, agents act without a mechanistic rule.…”
Section: Scientific Reportsmentioning
confidence: 99%
“…Rather, they consider every possible action and penalize those which are not favorable to their internal standards. While both the aforementioned model and LEUP are defined in a similar spirit, modeling under LEUP consists in correctly identifying the relevant internal cellular states for entropy optimization, while in 41 modeling is concerned with defining suitable penalizations for each possible decision scenario.…”
Section: Scientific Reportsmentioning
confidence: 99%