2016 International Joint Conference on Neural Networks (IJCNN) 2016
DOI: 10.1109/ijcnn.2016.7727386
|View full text |Cite
|
Sign up to set email alerts
|

PointNet: A 3D Convolutional Neural Network for real-time object class recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
92
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 157 publications
(93 citation statements)
references
References 12 publications
1
92
0
Order By: Relevance
“…Qi et al [25] gives a comprehensive study on the voxel-based CNNs and multi-view CNNs for 3D object classification. Other than those above, pointbased approach [11,24,15] is recently drawing much attention; however, the performance on 3D object classification is yet inferior to those of multi-view approaches. The current state-of-the-art result on the ModelNet40 benchmark dataset is reported by Wang et al [37], which is also based on the multi-view approach.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Qi et al [25] gives a comprehensive study on the voxel-based CNNs and multi-view CNNs for 3D object classification. Other than those above, pointbased approach [11,24,15] is recently drawing much attention; however, the performance on 3D object classification is yet inferior to those of multi-view approaches. The current state-of-the-art result on the ModelNet40 benchmark dataset is reported by Wang et al [37], which is also based on the multi-view approach.…”
Section: Related Workmentioning
confidence: 99%
“…Here, v i indicates the ID of the vertex where the i-th image of the object instance is observed. For instance, {v i } 20 i=1 in Candidate #2 is {1, 5,2,6,3,7,4,8,13,15,14,16,17,18,19,20,9,11,10, 12} ( Fig. 15 (b)).…”
Section: Sensitivity To Pre-defined Views Assumptionmentioning
confidence: 99%
“…Choosing what feature to use can be a difficult hyper-parameter tuning task. The other one is the network created feature by processing all points inside the cell, such as the PointNet structure [27] called voxel feature encoding (VFE) layer in VoxelNet. Each cell possesses a 128 dimensional feature learned by the network.…”
Section: Bird's Eye View Lidar Representationmentioning
confidence: 99%
“…This success has led researchers to apply similar methodology to 3D recognition tasks [46], facilitated by recent advances in computing that enable such tasks to be performed at scale. Seminal 3D classification datasets and efforts include ObjectNet3D [47], ShapeNet [48], VoxNet [49], and PointNet [50]. Most of these approaches focus on recognizing or creating objects with a given form and category (e.g., [51,52]), but there has been little work that seeks to derive the deeper relationship between desired functionality (e.g., performance and manufacturability) and requisite form (e.g.…”
Section: Machine Learning To Predict Am Qualitymentioning
confidence: 99%