2017 IEEE International Conference on Robotics and Automation (ICRA) 2017
DOI: 10.1109/icra.2017.7989538
|View full text |Cite
|
Sign up to set email alerts
|

SemanticFusion: Dense 3D semantic mapping with convolutional neural networks

Abstract: Ever more robust, accurate and detailed mapping using visual sensing has proven to be an enabling factor for mobile robots across a wide variety of applications. For the next level of robot intelligence and intuitive user interaction, maps need extend beyond geometry and appearence -they need to contain semantics. We address this challenge by combining Convolutional Neural Networks (CNNs) and a state of the art dense Simultaneous Localisation and Mapping (SLAM) system, ElasticFusion, which provides long-term d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
521
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 583 publications
(521 citation statements)
references
References 24 publications
0
521
0
Order By: Relevance
“…We solve this issue by utilizing the dense pixel‐wise semantic probability distribution produced by the neural network over every class. Therefore, we can improve the fusion accuracy by projecting the labels with a statistical approach using the Bayesian fusion [ASZ*16] [HFL14] [ZSS17] [MHDL17]. Bayesian fusion enables us to update the label prediction li on 2D images Ik within the common coordinate frame of the 3D model: Px=li|I1,,k=1ZPx=li|I1,,k1Px=li|Ik,…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We solve this issue by utilizing the dense pixel‐wise semantic probability distribution produced by the neural network over every class. Therefore, we can improve the fusion accuracy by projecting the labels with a statistical approach using the Bayesian fusion [ASZ*16] [HFL14] [ZSS17] [MHDL17]. Bayesian fusion enables us to update the label prediction li on 2D images Ik within the common coordinate frame of the 3D model: Px=li|I1,,k=1ZPx=li|I1,,k1Px=li|Ik,…”
Section: Methodsmentioning
confidence: 99%
“…Combined with SLAM systems, 2D semantic segmentation can be achieved in 3D environments [RA17] [TTLN17] [ZSS17] [MHDL17], a promising future in robotic vision understanding and autonomous driving. Unlike these existing methods that aimed at providing the semantic understanding of the scene for robots, we are focusing our attention on human interactions.…”
Section: Previous Workmentioning
confidence: 99%
“…There has been a great interest from the computer vision and robotics communities to exploit object-level information since from the perspective of many applications, it is beneficial to explore the awareness that object instances can provide for assistive computer vision [7,8,9], tracking/SLAM [10,11], or place categorization/scene recognition and life-long mapping [12,13].…”
Section: Related Workmentioning
confidence: 99%
“…For instance, Nascimento et al [30] applied a binary RGB-D descriptor to feed an Adaboost learning method to classify objects in a navigation task. McCormac et al [11] proposed a method for semantic 3D mapping. Their work combined the formulation of Whelan et al [31], an RGB-D based SLAM system for building a dense point cloud of the scene, with an encoder-decoder convolutional network for pixel-wise semantic segmentation.…”
Section: Slam and Augmented Semantic Representationsmentioning
confidence: 99%
“…The Dense Planar SLAM system [4] seamlessly maps an environment using planar and non-planar regions while tracking the sensor pose in real-time. SemanticFusion [5] combines Convolutional Neural Networks (CNNs) and a state-of-the-art dense SLAM system, ElasticFusion to create a dense semantic map. All of the four semantic SLAM systems mentioned above are focusing on the dense map, though semantic SLAM on the sparse map is fairly enough for some applications in fact.…”
Section: Introductionmentioning
confidence: 99%