2019
DOI: 10.1007/978-3-030-36150-1_18
|View full text |Cite
|
Sign up to set email alerts
|

Effects of a Social Force Model Reward in Robot Navigation Based on Deep Reinforcement Learning

Abstract: In this paper is proposed an inclusion of the Social Force Model (SFM) into a concrete Deep Reinforcement Learning (RL) framework for robot navigation. These types of techniques have demonstrated to be useful to deal with different types of environments to achieve a goal. In Deep RL, a description of the world to describe the states and a reward adapted to the environment are crucial elements to get the desire behaviour and achieve a high performance. For this reason, this work adds a dense reward function bas… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…In [34], the moving agents are simulated using the Optimal Reciprocal Collision Avoidance algorithm (ORCA) [36], and Imitation Learning is used to improve the performance. Although, as is described in [37], the performance is only high when the robot is close to the goal with few moving obstacles around.…”
Section: Deep Reinforcement Learning In Robot Navigationmentioning
confidence: 96%
“…In [34], the moving agents are simulated using the Optimal Reciprocal Collision Avoidance algorithm (ORCA) [36], and Imitation Learning is used to improve the performance. Although, as is described in [37], the performance is only high when the robot is close to the goal with few moving obstacles around.…”
Section: Deep Reinforcement Learning In Robot Navigationmentioning
confidence: 96%
“…For example, Yue et al [198] integrated SFM and a deep neural network in their Neural Social Physics model with learnable parameters. Gil and Sanfeliu [199] presented Social Force Generative Adversarial Network (SoFGAN) that uses a GAN and SFM to generate different plausible people trajectories reducing collisions in a scene.…”
Section: Human Trajectory Predictionmentioning
confidence: 99%
“…They receive a higher collaboration reward r co for not staying too close to each other. 25 Denote c 1 and c 2 as distance thresholds; and k another agent. The collaboration reward is computed as follows:…”
Section: Preliminariesmentioning
confidence: 99%