2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01289
|View full text |Cite
|
Sign up to set email alerts
|

Neural Topological SLAM for Visual Navigation

Abstract: This paper studies the problem of image-goal navigation which involves navigating to the location indicated by a goal image in a novel previously unseen environment. To tackle this problem, we design topological representations for space that effectively leverage semantics and afford approximate geometric reasoning. At the heart of our representations are nodes with associated semantic features, that are interconnected using coarse geometric information. We describe supervised learning-based algorithms that ca… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
182
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
1

Relationship

2
7

Authors

Journals

citations
Cited by 161 publications
(182 citation statements)
references
References 35 publications
0
182
0
Order By: Relevance
“…Oculus Insight, by Facebook, uses visual-inertial SLAM (simultaneous localization and mapping) to generate real-time maps and position tracking. 48 More sophisticated approaches, such as Neural Topological SLAM, leverage semantics and geometric information to improve long-horizon navigation (Chaplot et al 2020). Combining audio and visual sensors can further improve navigation of egocentric observations in complex 3D environments, which can be done through deep reinforcement learning approach (Chen et al 2020).…”
Section: Augmented Virtual and Mixed Reality (Vr Ar Mr)mentioning
confidence: 99%
“…Oculus Insight, by Facebook, uses visual-inertial SLAM (simultaneous localization and mapping) to generate real-time maps and position tracking. 48 More sophisticated approaches, such as Neural Topological SLAM, leverage semantics and geometric information to improve long-horizon navigation (Chaplot et al 2020). Combining audio and visual sensors can further improve navigation of egocentric observations in complex 3D environments, which can be done through deep reinforcement learning approach (Chen et al 2020).…”
Section: Augmented Virtual and Mixed Reality (Vr Ar Mr)mentioning
confidence: 99%
“…However, local matching localization using image level is still not sufficiently robust when the environment changes. Moreover, due to the widespread successful application of deep learning, the learning-based approach has also attracted a lot of interest [12] [13]. It relies on a large number of labeled datasets and cannot be applied well to an unfamiliar environment.…”
Section: Related Workmentioning
confidence: 99%
“…Planning and navigation in unstructured environments can be greatly improved by learning environment traversability using prior experience, but previous approaches to explicitly representing environment traversability require expensive human supervision or traversability heuristics [22][23][24][25]. Recent progress in using topological graphs to implicitly reason about traversability [18,[26][27][28] gives a promising way to learn from prior experience but has not been demonstrated for long-range navigation in complex, unstructured environments. In contrast, our hybrid approach employs a form of self-supervised learningbased explicit trajectory traversability estimation in tandem with geometry-based positional traversability estimation, which enables our approach to more robustly deal with previouslyunseen complex obstacles and terrains.…”
Section: Instantiating the Systemmentioning
confidence: 99%