2018
DOI: 10.48550/arxiv.1803.07289
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Flex-Convolution (Million-Scale Point-Cloud Learning Beyond Grid-Worlds)

Fabian Groh,
Patrick Wieschollek,
Hendrik P. A. Lensch

Abstract: Traditional convolution layers are specifically designed to exploit the natural data representation of images -a fixed and regular grid. However, unstructured data like 3D point clouds containing irregular neighborhoods constantly breaks the grid-based data assumption. Therefore applying best-practices and design choices from 2D-image learning methods towards processing point clouds are not readily possible. In this work, we introduce a natural generalization flex-convolution of the conventional convolution la… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(14 citation statements)
references
References 20 publications
0
14
0
Order By: Relevance
“…Many networks [5], [8], [40] that respect the permutationinvariant property are focused on designing convolution kernels for unordered 3D points. Previous studies [6], [7], [8], [43], [10] reveal effectiveness of considering local geometric details.…”
Section: F Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Many networks [5], [8], [40] that respect the permutationinvariant property are focused on designing convolution kernels for unordered 3D points. Previous studies [6], [7], [8], [43], [10] reveal effectiveness of considering local geometric details.…”
Section: F Discussionmentioning
confidence: 99%
“…Cires ¸an et al [38] replace subsampling layers with non-overlapping max-pooling layers in the CNNs [39], which achieves surprisingly rapid learning speed and better performance. Various subsampling methods on 3D point clouds [40], [6] have been proposed. Nonetheless, they either do not summarize local geometrical features, or ignore the problem caused by overlapped local neighborhood graphs.…”
Section: B Edge-preserved Poolingmentioning
confidence: 99%
“…However, conventional convolutions cannot be directly applied to point clouds due to the absence of regular grids. Previous networks mostly exploit local point features by two operations: local pooling [24,18,16] and flexible convolution [4,22,13,26]. Self-attention often uses linear layers, such as fully-connected (FC) layers and shared multilayer perceptron (shared MLP) layers, which are appropriate for point clouds.…”
Section: Related Workmentioning
confidence: 99%
“…Point-based Raw point clouds can be taken as input directly without converting them to another regular structured format. It is drawing more and more attention as it enjoys full sparsity [29,26,28,33,34,14,10,39]. Point-Net [26,29] proposes to use shared multi-layer perceptrons and max pooling layers to extract features of point clouds.…”
Section: Related Workmentioning
confidence: 99%
“…Methods OA(%) mAcc(%) Kd-Net [15] 77.4 82.3 SO-Net [19] 81.0 84.9 PCNN by Ext [2] 81.8 85.1 PointNet++ [28] 81.9 85.1 DGCNN [42] 82.3 85.1 SPLATNet [33] 83.7 85.4 Point Convolution KPConv [37] 85.0 86.2 SynSpecCNN [48] 82.0 84.7 Spidercnn [45] 82.4 85.3 SubSparseCNN [8] 83.3 86.0 PointCNN [20] 84.6 86.1 FlexConv [10] 85.0 84.7 PointNet++• P 82.0 85.4 KPConv• P2…”
Section: D Scene Segmentationmentioning
confidence: 99%