2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2019
DOI: 10.1109/iros40897.2019.8968170
|View full text |Cite
|
Sign up to set email alerts
|

3D LiDAR and Stereo Fusion using Stereo Matching Network with Conditional Cost Volume Normalization

Abstract: The complementary characteristics of active and passive depth sensing techniques motivate the fusion of the Li-DAR sensor and stereo camera for improved depth perception. Instead of directly fusing estimated depths across LiDAR and stereo modalities, we take advantages of the stereo matching network with two enhanced techniques: Input Fusion and Conditional Cost Volume Normalization (CCVNorm) on the LiDAR information. The proposed framework is generic and closely integrated with the cost volume component that … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
50
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 37 publications
(51 citation statements)
references
References 25 publications
1
50
0
Order By: Relevance
“…Currently, we implemented and tested our framework using the KITTI dataset (Geiger (2012)), and achieve a root mean squared error (RMSE) of 916.3 mm, which is better than the baseline (Kendall et al (2017)), and close to the state-of-art (Wang et al (2019)). In the future, we plan to improve our framework by adding more convolutional layers to our depth prediction network, implementing better cost functions to find the local optimum more accurately, and designing better refinement methods that take the disparities of neighbor points into consideration…”
Section: Current Results and Future Workmentioning
confidence: 90%
See 1 more Smart Citation
“…Currently, we implemented and tested our framework using the KITTI dataset (Geiger (2012)), and achieve a root mean squared error (RMSE) of 916.3 mm, which is better than the baseline (Kendall et al (2017)), and close to the state-of-art (Wang et al (2019)). In the future, we plan to improve our framework by adding more convolutional layers to our depth prediction network, implementing better cost functions to find the local optimum more accurately, and designing better refinement methods that take the disparities of neighbor points into consideration…”
Section: Current Results and Future Workmentioning
confidence: 90%
“…Previously, the general idea for this task is to plug LiDAR data into existing stereo matching algorithms. For example, Wang et al (2019) enhance the GC-Net (Kendall et al (2017)) by extracting more features and regularizing the cost volume using LiDAR data, while Park, Kim, and Sohn (2018) design a deep learning network to refine the output of SGM (Hirschmuller (2008)) with the help of LiDAR data. However, we argue that these stereo matching-base algorithms have two intrinsic drawbacks: first, intuitively, Copyright c 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org).…”
Section: Introductionmentioning
confidence: 99%
“…There have been very few attempts to fuse LiDAR and stereo in autonomous driving. Previous methods [16,17,18] were designed for the depth completion and, as far as we know, there is no fusion involved for 3D object detection benchmark dataset. Cheng et al The approach proposed here is specifically intended for the depth completion task and tested on a 3D object detection task.…”
Section: Related Workmentioning
confidence: 99%
“…To enrich the representation for a normal stereo (RGBs) matching network, a decision has been taken to join the geometry information from the LiDAR point cloud. However, instead of directly using a 3D point cloud from LiDAR, like in [18], the 4-beam LiDAR point cloud is reprojected to both left and right image coordinates using the calibration parameters to obtain the two sparse 4-beam LiDAR depth maps corresponding to stereo images. Unlike [18] that used a simple early fusion paradigm by concating stereo images with their corresponding sparse LiDAR depth maps, the proposed method uses a late fusion approach, presented in the following section.…”
Section: Inputmentioning
confidence: 99%
See 1 more Smart Citation