2018 IEEE International Conference on Robotics and Automation (ICRA) 2018
DOI: 10.1109/icra.2018.8461232
|View full text |Cite
|
Sign up to set email alerts
|

A General Pipeline for 3D Detection of Vehicles

Abstract: Autonomous driving requires 3D perception of vehicles and other objects in the in environment. Much of the current methods support 2D vehicle detection. This paper proposes a flexible pipeline to adopt any 2D detection network and fuse it with a 3D point cloud to generate 3D information with minimum changes of the 2D detection networks. To identify the 3D box, an effective model fitting algorithm is developed based on generalised car models and score maps. A two-stage convolutional neural network (CNN) is prop… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
90
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 152 publications
(91 citation statements)
references
References 38 publications
1
90
0
Order By: Relevance
“…LiDAR 3D Detection: The use of LiDAR data has proven to be essential input for SOTA frameworks [9,12,13,24,31,35,41] for 3D object detection applied to urban scenes.…”
Section: Related Workmentioning
confidence: 99%
“…LiDAR 3D Detection: The use of LiDAR data has proven to be essential input for SOTA frameworks [9,12,13,24,31,35,41] for 3D object detection applied to urban scenes.…”
Section: Related Workmentioning
confidence: 99%
“…The existing camera and LiDAR fusion methods can be divided into three different groups: 2D to 3D, proposal fusion, and dense fusion. In 2D to 3D approaches [5,19,21,30], 2D object detection is first performed on the RGB images using methods such as [17,23]. Afterwards, these 2D boxes are converted to 3D boxes using the LiDAR data.…”
Section: Object Detectionmentioning
confidence: 99%
“…Several methods for object detection in autonomous driving scenes use lidar sensor data alongside images from an aligned camera [10], [11], [17], using the KITTI data set for training and evaluation [13]. While the KITTI data set is an excellent benchmark, its annotations are 3D bounding boxes, and object classes are limited to driving-relevant objects such as pedestrians, cars, and cyclists.…”
Section: A Deep Learning On Point Cloudsmentioning
confidence: 99%
“…The camera provides rich semantic information about object classes and facilitates the application of deep neural networks, and the lidar sensor provides precise 3D spatial measurements. Additionally, by identifying objects in 3D at the point level, rather than outputting rectangular 3D bounding boxes around detected objects [9], [16], [10], [11], [17], LDLS allows much more precise localization of object instances which do not neatly fit into rectangular boxes, such as people, animals, etc.…”
mentioning
confidence: 99%