2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2019
DOI: 10.1109/iros40897.2019.8968242
|View full text |Cite
|
Sign up to set email alerts
|

Improving 3D Object Detection for Pedestrians with Virtual Multi-View Synthesis Orientation Estimation

Abstract: Accurately estimating the orientation of pedestrians is an important and challenging task for autonomous driving because this information is essential for tracking and predicting pedestrian behavior. This paper presents a flexible Virtual Multi-View Synthesis module that can be adopted into 3D object detection methods to improve orientation estimation. The module uses a multi-step process to acquire the finegrained semantic information required for accurate orientation estimation. First, the scene's point clou… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2020
2020
2025
2025

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 27 publications
(3 citation statements)
references
References 43 publications
(82 reference statements)
0
3
0
Order By: Relevance
“…Sparse Dd point cloud data is encoded with compact multi-view. Ku used a virtual multi-view synthesis method to generate a set of virtual views for each detected pedestrian [ 19 ]. These views are used to create accurate position estimates during training and reasoning.…”
Section: Realted Workmentioning
confidence: 99%
“…Sparse Dd point cloud data is encoded with compact multi-view. Ku used a virtual multi-view synthesis method to generate a set of virtual views for each detected pedestrian [ 19 ]. These views are used to create accurate position estimates during training and reasoning.…”
Section: Realted Workmentioning
confidence: 99%
“…MMF [20] benefits from multi-task learning and multi-sensor fusion. VMVS [16] generates a set of virtual views for each detected pedestrian in pseudo point clouds. Then the different views are used to produce an accurate orientation estimation.…”
Section: Related Workmentioning
confidence: 99%
“…As can be seen in Figure 2, the performance of approaches which have a focus on LiDAR for the final predictions (F-PointNet and PointPillars) have a significant drop in performance for long range when compared to approaches like AVOD which use both sensors. VMVS [13], STD [14], IPOD [12], F-ConvNet [15] and TANet [16] were not evaluated in detail, due to the lack of source code available or similarity to other approaches. VMVS, F-ConvNet and TANet are conceptually similar to AVOD, F-PointNet or PointPillars.…”
Section: Introductionmentioning
confidence: 99%