2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01597
|View full text |Cite
|
Sign up to set email alerts
|

Perception-Aware Multi-Sensor Fusion for 3D LiDAR Semantic Segmentation

Abstract: 3D LiDAR (light detection and ranging) based semantic segmentation is important in scene understanding for many applications, such as auto-driving and robotics. For example, for autonomous cars equipped with RGB cameras and LiDAR, it is crucial to fuse complementary information from different sensors for robust and accurate segmentation. Existing fusion-based methods, however, may not achieve promising performance due to the vast difference between two modalities. In this work, we investigate a collaborative f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
45
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 134 publications
(54 citation statements)
references
References 54 publications
0
45
0
Order By: Relevance
“…Polarimetric optical cues [11], [82], [83] and event-driven priors [12], [84], [85] are often intertwined for robust perception under adverse conditions. In automated driving, optical flow [86] and LiDAR data [87] are also incorporated for enhanced semantic road scene understanding. However, most of these works only address a single sensing-modality combination scenario.…”
Section: Multi-modal Semantic Segmentationmentioning
confidence: 99%
“…Polarimetric optical cues [11], [82], [83] and event-driven priors [12], [84], [85] are often intertwined for robust perception under adverse conditions. In automated driving, optical flow [86] and LiDAR data [87] are also incorporated for enhanced semantic road scene understanding. However, most of these works only address a single sensing-modality combination scenario.…”
Section: Multi-modal Semantic Segmentationmentioning
confidence: 99%
“…monocular RGB images and point clouds. This is also inspired by multi-sensor-based methods on object tracking [55] , object detection [30,36] and point cloud segmentation [22,56]. In addition, we detect noisy labels and propose a noisy-label-aware training scheme to reduce the negative effect of noisy labels on the training, different from existing methods.…”
Section: Related Workmentioning
confidence: 99%
“…Caltagirone et al [ 25 ] developed a novel multimodal system for road detection by fusing LiDAR and camera data, and the cross function convolutional neural network (FCN) architecture was introduced. Zhuang et al [ 26 ] investigated a collaborative fusion scheme called perception-aware multi-sensor fusion to exploit perceptual information from two modalities that included appearance information from RGB images and the spatio-depth information from point clouds, which achieved the 3D semantic segmentation. An offline LiDAR and camera fusion method was proposed by Zhen et al [ 27 ] to build dense and accurate 3D models.…”
Section: Related Workmentioning
confidence: 99%