2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00469
|View full text |Cite
|
Sign up to set email alerts
|

PVNet: Pixel-Wise Voting Network for 6DoF Pose Estimation

Abstract: This paper addresses the challenge of 6DoF pose estimation from a single RGB image under severe occlusion or truncation. Many recent works have shown that a two-stage approach, which first detects keypoints and then solves a Perspective-n-Point (PnP) problem for pose estimation, achieves remarkable performance. However, most of these methods only localize a set of sparse keypoints by regressing their image coordinates or heatmaps, which are sensitive to occlusion and truncation. Instead, we introduce a Pixel-w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

2
921
2

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 920 publications
(925 citation statements)
references
References 46 publications
2
921
2
Order By: Relevance
“…Under the KPEC setting, we developed a monocular pose estimation technique for space-borne objects such as satellites. Inspired by works that combine the strength of deep neural networks and geometric optimisation [26,25,35], our approach contains three main components:…”
Section: Introductionmentioning
confidence: 99%
“…Under the KPEC setting, we developed a monocular pose estimation technique for space-borne objects such as satellites. Inspired by works that combine the strength of deep neural networks and geometric optimisation [26,25,35], our approach contains three main components:…”
Section: Introductionmentioning
confidence: 99%
“…Existing methods for localising objects in 3D or estimating their 6 DoF pose rely on databases of 3D object models [5,6,7,8] or need motion capture systems with markers [4,9,10]. To avoid using markers, feature points [11,12] can be localised in an image and matched against a 3D object model to estimate the object pose by solving a Perspective-n-Point (PnP) problem [13].…”
Section: Introductionmentioning
confidence: 99%
“…For example, DenseFusion [5] combines features obtained from RGB-D images and can handle occlusions and inaccurate segmentation. Pixel-wise Voting Network (PVNet) [6] estimates the pose of occluded or truncated objects with an uncertainty-driven PnP, learning a vector-field representation to localise a sparse set of 2D keypoints and their spatial uncertainty. Normalized Object Coordinate Space (NOCS) [16] uses a normalised object coordinates space formulation that jointly estimates the 6 DoF pose and the dimensions (in the form of a 3D bounding box) of a novel object (i.e.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations