2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2022
DOI: 10.1109/wacv51458.2022.00076
|View full text |Cite
|
Sign up to set email alerts
|

Data Augmented 3D Semantic Scene Completion with 2D Segmentation Priors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 31 publications
0
5
0
Order By: Relevance
“…Semantic scene completion is achieved in [33][34][35][36][37][38][39][40][41][42][43][44][45][46][47][48][49][50] by adopting convolutions, transformers, or a combination of both, on different modalities like RGB images, depth images, and truncated signed distance functions (TSDFs), where the distance to the closest TSDF surface is computed at given 3D locations (usually voxel centers). Liu et al [33] proposed a disentangled framework, sequentially carrying out 2D semantic segmentation, 2D-3D projection, and 3D semantic scene completion.…”
Section: Multi-modality-based Methodsmentioning
confidence: 99%
“…Semantic scene completion is achieved in [33][34][35][36][37][38][39][40][41][42][43][44][45][46][47][48][49][50] by adopting convolutions, transformers, or a combination of both, on different modalities like RGB images, depth images, and truncated signed distance functions (TSDFs), where the distance to the closest TSDF surface is computed at given 3D locations (usually voxel centers). Liu et al [33] proposed a disentangled framework, sequentially carrying out 2D semantic segmentation, 2D-3D projection, and 3D semantic scene completion.…”
Section: Multi-modality-based Methodsmentioning
confidence: 99%
“…A common requirement in applications is to infer semantic labels not only in directly observed but also in occluded space; to this end, semantic scene completion [34] seeks to address both scene occupancy completion and semantic object labeling jointly. Single-view depth images can be viewed as minimal input data [18, 34-36, 41, 42]; alternatives [2,12,14,22] (includ-ing our method) tackle completing fused reconstructions of entire 3D spaces. [12] jointly predicts a truncated, unsigned distance field and per-volume semantics in a series of hierarchy levels ranging from low to high resolution.…”
Section: Related Workmentioning
confidence: 99%
“…[12] jointly predicts a truncated, unsigned distance field and per-volume semantics in a series of hierarchy levels ranging from low to high resolution. Leveraging RGB-D image back-projection, [22] combines geometry and appearance features to infer semantics and refine 3D geometry; [14] adopts a view of RGB image segmentation as a prior and computes the final semantic scene completion using a 3D CNN. [2] exploits an interplay between the scene-and instance-level completion tasks and alternates between semantic scene completion and detection of object instances.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Later, some followers (Guo and Tong 2018;Liu et al 2018;Garbade et al 2019;Li et al 2020b) leveraged the 2D semantic priors of color images via feature projection to improve the performance. Recent works (Dourado et al 2021;Dourado, Guth, and de Campos 2022;Wang, Lin, and Wan 2022) further fused complex features that are extracted from depth or color images to 2D semantic network. In addition, IMENet (Li, Ding, and Huang 2021) proposed an iterative fusion scheme to ensure the branches of 2D segmentation and 3D scene completion fully benefit each other.…”
Section: Related Workmentioning
confidence: 99%