Proceedings of the 10th International Conference on Computer Vision Theory and Applications 2015
DOI: 10.5220/0005299304860490
|View full text |Cite
|
Sign up to set email alerts
|

Learning Visual Odometry with a Convolutional Network

Abstract: Abstract:We present an approach to predicting velocity and direction changes from visual information ("visual odometry") using an end-to-end, deep learning-based architecture. The architecture uses a single type of computational module and learning rule to extract visual motion, depth, and finally odometry information from the raw data. Representations of depth and motion are extracted by detecting synchrony across time and stereo channels using network layers with multiplicative interactions. The extracted re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
88
0
2

Year Published

2017
2017
2021
2021

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 140 publications
(90 citation statements)
references
References 7 publications
0
88
0
2
Order By: Relevance
“…Numerous approaches [1,14,21,24,25] learn the task of VO estimation using ground-truth data available in the form of global-camera poses, recorded by high-precision GPU+IMS rigs.…”
Section: Supervised Approachesmentioning
confidence: 99%
See 2 more Smart Citations
“…Numerous approaches [1,14,21,24,25] learn the task of VO estimation using ground-truth data available in the form of global-camera poses, recorded by high-precision GPU+IMS rigs.…”
Section: Supervised Approachesmentioning
confidence: 99%
“…Konda et.al. [14] first proposed an autoencoder to learn a latent representation of the optical flow between camera frames jointly with the ego-motion estimation task. Kendall et.al.…”
Section: Supervised Approachesmentioning
confidence: 99%
See 1 more Smart Citation
“…This allows the freedom to model various unconstrained and partially constrained motions that typically affect the overall robustness of existing ego-motion algorithms. While model-based approaches have shown tremendous progress in accuracy, robustness, and run-time performance, a few recent data-driven approaches have been shown to produce equally compelling results [20], [22], [24]. An adaptive and trainable solution for relative pose estimation or ego-motion can be especially advantageous for several reasons: (i) a generalpurpose end-to-end trainable model architecture that applies to a variety of camera optics including pinhole, fisheye, and catadioptric lenses; (ii) simultaneous and continuous optimization over both ego-motion estimation and camera parameters (intrinsics and extrinsics that are implicitly modeled); and (iii) joint reasoning over resource-aware computation and accuracy within the same architecture is amenable.…”
Section: Ego-motion Regressionmentioning
confidence: 99%
“…CNNs are also capable of ego-motion estimation (Konda and Memisevic, 2015). However, the results need to be improved to compete with conventional methods.…”
Section: Introductionmentioning
confidence: 99%