2021
DOI: 10.3390/s21144813
|View full text |Cite
|
Sign up to set email alerts
|

Lidar–Camera Semi-Supervised Learning for Semantic Segmentation

Abstract: In this work, we investigated two issues: (1) How the fusion of lidar and camera data can improve semantic segmentation performance compared with the individual sensor modalities in a supervised learning context; and (2) How fusion can also be leveraged for semi-supervised learning in order to further improve performance and to adapt to new domains without requiring any additional labelled data. A comparative study was carried out by providing an experimental evaluation on networks trained in different setups … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3
1
1

Relationship

2
8

Authors

Journals

citations
Cited by 15 publications
(9 citation statements)
references
References 25 publications
0
9
0
Order By: Relevance
“…Camera data is fused with LiDAR data in order to detect better objects [ 26 ]. In some works, the detection of objects is approached by performing semantic segmentation on LiDAR data [ 29 , 30 ] or camera-LiDAR fused data [ 31 ].…”
Section: Related Workmentioning
confidence: 99%
“…Camera data is fused with LiDAR data in order to detect better objects [ 26 ]. In some works, the detection of objects is approached by performing semantic segmentation on LiDAR data [ 29 , 30 ] or camera-LiDAR fused data [ 31 ].…”
Section: Related Workmentioning
confidence: 99%
“…Besides, different Lidars can work cooperatively to better monitor traffic situations (Kloeker et al, 2020). Even when Lidars are fused with radars or cameras, Lidars are the fundamental sensor for positioning traffic agents (Caltagirone et al, 2019). Therefore, this study emphasizes Lidar sensors as RSUs.…”
Section: Introductionmentioning
confidence: 99%
“…More importantly, latefusion systems incorporate single-modality detectors. Therefore, our method projects point clouds into the camera plane to create a three-channel tensor with the same width and height of the image, of which each channel encodes one of the 3D spatial coordinates [9].…”
Section: Introductionmentioning
confidence: 99%