2023
DOI: 10.1109/tmm.2022.3189778
|View full text |Cite
|
Sign up to set email alerts
|

VPFNet: Improving 3D Object Detection With Virtual Point Based LiDAR and Stereo Data Fusion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 85 publications
(15 citation statements)
references
References 59 publications
0
14
0
1
Order By: Relevance
“…In order to construct a more robust system, there is an urgent need to collect collaborative perception data in complex environments and propose well-designed methods for various complex scenarios. Multi-sensor fusion helps compensate for weather and distance's effects on data quality, and virtual point cloud generation [108,109,110] can be used to predict long-range objects. Additionally, spatio-temporal data fusion is required to predict the trajectory of objects moving at high speeds, and these are promising directions for future research.…”
Section: B Collaboration Perception In Complex Scenesmentioning
confidence: 99%
“…In order to construct a more robust system, there is an urgent need to collect collaborative perception data in complex environments and propose well-designed methods for various complex scenarios. Multi-sensor fusion helps compensate for weather and distance's effects on data quality, and virtual point cloud generation [108,109,110] can be used to predict long-range objects. Additionally, spatio-temporal data fusion is required to predict the trajectory of objects moving at high speeds, and these are promising directions for future research.…”
Section: B Collaboration Perception In Complex Scenesmentioning
confidence: 99%
“…Another solution based on LIDAR and multiview representations is proposed by [ 4 ]. They published a new architecture called VPFNet that aligns and aggregates the image data and point cloud at virtual points that can bridge the resolution gap between the two sensors (LIDAR and multiview images).…”
Section: Related Workmentioning
confidence: 99%
“…To address these problems, many attempts have been made in recent studies. Zhu et al [ 37 ] proposed VPFNet, a network that first generates 3D proposal regions from point clouds and divides them into multiple small grids again, with the corner points of each grid as virtual points, and next projects these points onto the image for image feature sampling and point cloud feature extraction, which reduces the computation and compensates the lack of sparse point clouds. Huang et al [ 20 ] proposed a LI-Fusion module for fusing point cloud and image information at different scales in the feature extraction stage, and then associating the image features to the corresponding point cloud locations, while adding a consistency forcing loss for balancing the impact of classification loss and location loss on the detection results.…”
Section: Related Workmentioning
confidence: 99%
“…In recent years, there are some studies on 3D object detection based on camera and LiDAR fusion, which can be classified into serial type [ 10 , 11 , 12 , 13 , 14 , 15 , 16 ] and parallel type [ 17 , 18 , 19 , 20 , 21 , 22 , 23 , 24 , 25 , 26 , 27 , 28 , 29 , 30 , 31 , 32 , 33 , 34 , 35 , 36 , 37 , 38 ] according to the stage of fusion. The serial type method is represented by F-PointNet [ 13 ] which usually takes the image of the camera as input first and uses image object detection or semantic segmentation algorithm to get the spatial location of the object, then projects it to the LiDAR point cloud to extract the point cloud of the frustum region around the object, and finally uses the normal point cloud 3D object detection algorithm to get 3D bounding boxes.…”
Section: Introductionmentioning
confidence: 99%