2021
DOI: 10.3390/app11125598
|View full text |Cite
|
Sign up to set email alerts
|

Radar Voxel Fusion for 3D Object Detection

Abstract: Automotive traffic scenes are complex due to the variety of possible scenarios, objects, and weather conditions that need to be handled. In contrast to more constrained environments, such as automated underground trains, automotive perception systems cannot be tailored to a narrow field of specific tasks but must handle an ever-changing environment with unforeseen events. As currently no single sensor is able to reliably perceive all relevant activity in the surroundings, sensor data fusion is applied to perce… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 39 publications
(13 citation statements)
references
References 51 publications
0
13
0
Order By: Relevance
“…This operation causes information loss. However, works like [84] directly fused camera, LiDAR, and radar inputs without initial processing.…”
Section: A Data (Early) Fusionmentioning
confidence: 99%
See 2 more Smart Citations
“…This operation causes information loss. However, works like [84] directly fused camera, LiDAR, and radar inputs without initial processing.…”
Section: A Data (Early) Fusionmentioning
confidence: 99%
“…Method Dataset Input Fusion Type mAP-BEV/NDS (%) mAP-3D/mAP (%) MV3D [4] KITTI [32] RGB image& LiDAR early, middle, late --PointFusion [85] KITTI [32] RGB image& LiDAR early -40.13 AVOD-FPN [5] KITTI [32] RGB image &LiDAR middle 64.03 55.63 SAANET [111] KITTI [32] RGB imag & LiDAR middle -52.5 3D-CVF [58] KITTI [32] RGB image &LiDAR middle (gated based) --CenterFusion [109] nuScenes [30] RGB image & radar middle 45.30 33.20 RVF-Net [84] nuScenes [30] RGB image, radar, & LiDAR early 54.86 -FusionNet [96] custom [96] RGB image & radar early -73.5 Meyer et al [82] Astyx [40] RGB image & radar late -48.0 VPFNet [124] KITTI [32] RGB If we have a deeper network, we lose much information at each layer because of this operation. Coming up with a new idea to replace a pooling operation with an equivalent operation that does not cause information loss or build a reconstruction /upsampling layer that can fully reconstruct the lost features with no loss, such as using wavelets, may help to increase the detection performance.…”
Section: Other Fusion Techniquesmentioning
confidence: 99%
See 1 more Smart Citation
“…These point cloud based networks can be further differentiated into grid-based and point-based architectures. Grid-based approaches first render the point cloud into a 2D bird eye view (BEV) or 3D voxel grid using hand-crafted operations [11], [28], [29], [30], [31] or learned feature-encoders [32], [12], [31] and subsequently apply convolutional backbones to the grid.…”
Section: A Radar Object Detectionmentioning
confidence: 99%
“…In the field of target detection [35][36][37], recall and precision are mainly used as the performance measure of the algorithm. Precision (P) and recall (R) are, respectively, defined as follows: R = TP TP + FN ,…”
Section: Algorithm Performance Evaluationmentioning
confidence: 99%