2020 International Conference on 3D Vision (3DV) 2020
DOI: 10.1109/3dv50981.2020.00088
|View full text |Cite
|
Sign up to set email alerts
|

LiDAR-based Recurrent 3D Semantic Segmentation with Temporal Memory Alignment

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 31 publications
(13 citation statements)
references
References 21 publications
0
13
0
Order By: Relevance
“…For the quantitative evaluation, we compare our proposed approach against other methods on the multiple scans task in semantic segmentation benchmark. As shown in Table I, our presented Meta-RangeSeg method outperforms the stateof-the-art approach [26] by 2.5% on SemanticKITTI test set, which also achieves the best results on the most of classes. It is worthy of mentioning that our approach can process the input point cloud data at a rate of 22Hz while maintaining high accuracy.…”
Section: Performance Evaluationmentioning
confidence: 81%
See 1 more Smart Citation
“…For the quantitative evaluation, we compare our proposed approach against other methods on the multiple scans task in semantic segmentation benchmark. As shown in Table I, our presented Meta-RangeSeg method outperforms the stateof-the-art approach [26] by 2.5% on SemanticKITTI test set, which also achieves the best results on the most of classes. It is worthy of mentioning that our approach can process the input point cloud data at a rate of 22Hz while maintaining high accuracy.…”
Section: Performance Evaluationmentioning
confidence: 81%
“…Shi et al [14] employ a voxel-based 3D sparse convolutional network to fuse local information from the previous and current frames through local interpolation, which only make use of the two consecutive scans. Duerr et al [26] propose a recurrent segmentation architecture using range images, which recursively aggregate the features of previous scans in order to exploit the short term temporal dependencies. In [15], superimposing point clouds in 3D space is adopted for multiple scans segmentation, whose memory consumption and computational time increase linearly with the total number of scans per input model.…”
Section: Related Workmentioning
confidence: 99%
“…PandaSet [29] provides 6080 point-wise annotated lidar scans of a Pandar64, with corresponding camera images from the front center camera. For comparability with other approaches [23] we group the labeled classes into a subset of 14 classes and follow the data split proposed in [23].…”
Section: Discussionmentioning
confidence: 99%
“…The lidar backbone computes features for the input point clouds, which are represented in range view, based on the spherical projection of [21], [23]. Its architecture is motivated by EfficientPS [8] and adapted to the range view.…”
Section: A Lidar Backbonementioning
confidence: 99%
“…Previous works tackled this problem by extracting temporal information from residual range images [5] or bird's eye view (BEV) images [26], typically using a 2D convolutional neural network (CNN). The back-projection from these 2D representations to the 3D space often requires post-processing like k-nearest neighbor (kNN) clustering [5], [9], [12], [25] to avoid labels bleeding into points that are close in the image space but distant in 3D. Other approaches can identify objects that have moved in 3D space directly during mapping [1] or with a clustering and tracking approach [6].…”
Section: Non-moving Movingmentioning
confidence: 99%