2019
DOI: 10.1007/978-3-030-20873-8_44
|View full text |Cite
|
Sign up to set email alerts
|

SPNet: Deep 3D Object Classification and Retrieval Using Stereographic Projection

Abstract: We propose an efficient Stereographic Projection Neural Network (SPNet) for learning representations of 3D objects. We first transform a 3D input volume into a 2D planar image using stereographic projection. We then present a shallow 2D convolutional neural network (CNN) to estimate the object category followed by view ensemble, which combines the responses from multiple views of the object to further enhance the predictions. Specifically, the proposed approach consists of four stages: (1) Stereographic projec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
29
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 43 publications
(29 citation statements)
references
References 28 publications
0
29
0
Order By: Relevance
“…Over recent years, several DL approaches are emerging. In contrast to multiview-based [48][49][50][51][52] and voxel-based approaches [41,43,53,54], such approaches do not need specific pre-processing steps, and have been proved to provide state-of-the-art performances in semantic segmentation and classification task on standard benchmarks. Since the main objective of this work is to directly exploit the three-dimensionality of Point Clouds, a comprehensive overview of the multiview and voxel based methods is out of the scope, therefore only the point-based methods will be detailed.…”
Section: Semantic Segmentation Of Point Cloudsmentioning
confidence: 99%
“…Over recent years, several DL approaches are emerging. In contrast to multiview-based [48][49][50][51][52] and voxel-based approaches [41,43,53,54], such approaches do not need specific pre-processing steps, and have been proved to provide state-of-the-art performances in semantic segmentation and classification task on standard benchmarks. Since the main objective of this work is to directly exploit the three-dimensionality of Point Clouds, a comprehensive overview of the multiview and voxel based methods is out of the scope, therefore only the point-based methods will be detailed.…”
Section: Semantic Segmentation Of Point Cloudsmentioning
confidence: 99%
“…ModelNet10 Acc ModelNet40 Acc RotationNet [45] 98.46% 97.37% SPNet [46] 97.25% 92.63% SO-Net [18] 95,07% 93,40% Point2Sequence [47] 95,30% 92.60% Lonchanet [22] 94,37% -VFD 92.84% 88.74% binVoxNetPlus [48] 92.32% 85.47% Primitive-GAN [49] 92.20% 86.00% VSL [50] 91.00% 84.50% OrthographicNet [51] 88.56% -3DShapeNets [42] 83.50% 77.00% PointNet [43] 77,6% -…”
Section: Papermentioning
confidence: 99%
“…This engine employs 2D CNN to extract view features and matches them to calculate the similarity between 3D models. Seout et al [41] propose a stereographic projection neural network (SPNet). This method learns the feature representation of a 3D model by transforming the input 3D model into a 2D planar image using stereo projection.…”
Section: Related Workmentioning
confidence: 99%