2020
DOI: 10.3390/app10238386
|View full text |Cite
|
Sign up to set email alerts
|

A Deep Reinforcement Learning Approach for Active SLAM

Abstract: In this paper, we formulate the active SLAM paradigm in terms of model-free Deep Reinforcement Learning, embedding the traditional utility functions based on the Theory of Optimal Experimental Design in rewards, and therefore relaxing the intensive computations of classical approaches. We validate such formulation in a complex simulation environment, using a state-of-the-art deep Q-learning architecture with laser measurements as network inputs. Trained agents become capable not only to learn a policy to navig… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
26
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 30 publications
(26 citation statements)
references
References 49 publications
0
26
0
Order By: Relevance
“…However, note that decisions will be optimal only locally and a short decision-making horizon may induce wrong behaviors [27], [114]. This strategy is typical in deep reinforcement learning (DRL) approaches [59], [115], for which local optimality is alleviated by network memorization. Following the idea that evaluating larger neighborhoods would lead to more robust decisions, in [41] authors use RRT-based paths to several configurations over the free space as the action set; and in [44] the entire environment is considered under the umbrella of continuous-domain optimization.…”
Section: B Detecting Goal Locationsmentioning
confidence: 99%
See 2 more Smart Citations
“…However, note that decisions will be optimal only locally and a short decision-making horizon may induce wrong behaviors [27], [114]. This strategy is typical in deep reinforcement learning (DRL) approaches [59], [115], for which local optimality is alleviated by network memorization. Following the idea that evaluating larger neighborhoods would lead to more robust decisions, in [41] authors use RRT-based paths to several configurations over the free space as the action set; and in [44] the entire environment is considered under the umbrella of continuous-domain optimization.…”
Section: B Detecting Goal Locationsmentioning
confidence: 99%
“…Optimality criteria were first used in active SLAM by Feder et al [27], where utility was computed as the area of the covariance ellipses describing the uncertainty in the robot's and map's posteriors. Since then, many active SLAM methods based on TOED have been proposed, mostly based on T -opt [35], [40] and, recently, on D-opt [59], [60]. Even so, TOED criteria are not as popular as IT approaches.…”
Section: Theory Of Optimal Experimental Design (Toed)mentioning
confidence: 99%
See 1 more Smart Citation
“…Some methods [14], [15] formulate this problem as the POMDP (Partially Observable Markov Decision Process) [49]. Recently, some researchers have designed learningbased policies [17]- [24] to tackle this problem. Chaplot et al [18] propose a novel module that combines a hierarchical network design and classical path planning, which significantly improves the sample efficiency and leads to better performance.…”
Section: Related Work a Active Slammentioning
confidence: 99%
“…In (Botteghi et al, 2020) and (Placed and Castellanos, 2020), the authors propose a reinforcement learning approach for active SLAM that relies on 2D LiDAR readings and information coming from the SLAM algorithm to select the best steering command, out of a discrete action space, for a mobile robot to explore different environments. While we use similar state space definition, we employ a continuous action space for obtaining smoother trajectories.…”
Section: Reinforcement Learning-based Active Slam In Roboticsmentioning
confidence: 99%