2020
DOI: 10.1609/aaai.v34i07.6945
|View full text |Cite
|
Sign up to set email alerts
|

ZoomNet: Part-Aware Adaptive Zooming Neural Network for 3D Object Detection

Abstract: 3D object detection is an essential task in autonomous driving and robotics. Though great progress has been made, challenges remain in estimating 3D pose for distant and occluded objects. In this paper, we present a novel framework named ZoomNet for stereo imagery-based 3D detection. The pipeline of ZoomNet begins with an ordinary 2D object detection model which is used to obtain pairs of left-right bounding boxes. To further exploit the abundant texture cues in rgb images for more accurate disparity estimatio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
33
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 73 publications
(33 citation statements)
references
References 15 publications
0
33
0
Order By: Relevance
“…1) 2D-based methods [7,54,30,39,57,49,38] first detect 2D bounding box proposals and then regress instancewise 3D boxes. Stereo-RCNN [30] extended Faster R-CNN [43] for stereo-inputs to associate left and right images.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…1) 2D-based methods [7,54,30,39,57,49,38] first detect 2D bounding box proposals and then regress instancewise 3D boxes. Stereo-RCNN [30] extended Faster R-CNN [43] for stereo-inputs to associate left and right images.…”
Section: Related Workmentioning
confidence: 99%
“…Stereo-RCNN [30] extended Faster R-CNN [43] for stereo-inputs to associate left and right images. Disp R-CNN [49] and ZoomNet [57] incorporated extra instance segmentation mask and part location map to improve detection quality. However, the final performance is limited by the recall of 2D detection algorithms, and 3D geometry information is not fully utilized.…”
Section: Related Workmentioning
confidence: 99%
“…To draw and visualize a object's orientation from a bird's eye view perspective, the object's α will need to be converted into r y . With the object's x and z location information, we can convert r y to α or vice versa [40,41,42], using the formula:…”
Section: Alpha (Local/allocentric Rotation)mentioning
confidence: 99%
“…A robust autonomous driving system requires its LiDAR-based detector to reliably handle different environmental conditions, e.g., geographic locations and weather conditions. While 3D detection has received increasing interest in recent years, most existing works [79,7,10,11,16,23,25,26,27,29,30,36,41,42,49,50,51,53,63,64,65,67,68,62,78] The regions of "missing points" are irregular in shape.…”
Section: Introductionmentioning
confidence: 99%