2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2019
DOI: 10.1109/iros40897.2019.8967803
|View full text |Cite
|
Sign up to set email alerts
|

Deep Sensor Fusion for Real-Time Odometry Estimation

Abstract: Cameras and 2D laser scanners, in combination, are able to provide low-cost, light-weight and accurate solutions, which make their fusion well-suited for many robot navigation tasks. However, correct data fusion depends on precise calibration of the rigid body transform between the sensors. In this paper we present the first framework that makes use of Convolutional Neural Networks (CNNs) for odometry estimation fusing 2D laser scanners and mono-cameras. The use of CNNs provides the tools to not only extract t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 27 publications
0
7
0
Order By: Relevance
“…When provided with data from multiple sensors, a common approach in the literature is to fuse readings coming from different sensors together [28], [29]. To explore this option, we train the Sensor Fusion (SF) on the task of regressing the drone position from both image and audio features, using only instances in T .…”
Section: Alternative Strategiesmentioning
confidence: 99%
See 1 more Smart Citation
“…When provided with data from multiple sensors, a common approach in the literature is to fuse readings coming from different sensors together [28], [29]. To explore this option, we train the Sensor Fusion (SF) on the task of regressing the drone position from both image and audio features, using only instances in T .…”
Section: Alternative Strategiesmentioning
confidence: 99%
“…Most of the strategies considered in this work employ a CNN architecture based on MobileNet-V2 [31], with a total of 1 million parameters, and a variable number of output neurons dependant on the chosen strategy (3 for strictly supervised approaches, 12 for the SaP model). The SF model utilizes the same convolutional architecture for the image branch, while a series of 4 feed-forward layers with ReLU non-linearities processes the audio information and 3 more feed-forward layers fuse the two streams, similarly to [29]. The AaP model implements an encoder-decoder CNN architecture, with a bottleneck of size 128 [1].…”
Section: E Neural Network Architectures and Trainingmentioning
confidence: 99%
“…However, traditional methods need accurate models and careful calibration, so employing machine learning for sensor fusion in odometry has also become an open research topic. In [ 299 ], sequences of CNNs were used to extract features and determine pose from a camera and 2D LiDAR. Some learning-based methods, such as VINet and DeepVIO [ 300 ], demonstrate comparable or even better performance than traditional methods.…”
Section: Sensor Fusionmentioning
confidence: 99%
“…This data is then sequentially fed to a 1D convolution layer. We use six 1D convolutional layers following the method in [41]. Each convolutional layer is followed by a ReLU activation.…”
Section: Camera Feature Extractionmentioning
confidence: 99%