2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00643
|View full text |Cite
|
Sign up to set email alerts
|

SeGAN: Segmenting and Generating the Invisible

Abstract: Objects often occlude each other in scenes; Inferring their appearance beyond their visible parts plays an important role in scene understanding, depth estimation, object interaction and manipulation. In this paper, we study the challenging problem of completing the appearance of occluded objects. Doing so requires knowing which pixels to paint (segmenting the invisible parts of objects) and what color to paint them (generating the invisible parts). Our proposed novel solution, SeGAN, jointly optimizes for bot… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
142
2
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 168 publications
(145 citation statements)
references
References 51 publications
0
142
2
1
Order By: Relevance
“…Layered scene representations come in a diversity of contexts, such as depth ordering of semantic maps [48,39,16] and color images [7], motion analysis and optical flow [42,38], stereo reconstruction [3], scene decomposition in depth surfaces [28] and planes [27]. Our focus is on the Layered Depth Images (LDI) introduced by Shade et al [35], which refer to a single view representation of a scene that contains multiple layers of RGB-D information.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Layered scene representations come in a diversity of contexts, such as depth ordering of semantic maps [48,39,16] and color images [7], motion analysis and optical flow [42,38], stereo reconstruction [3], scene decomposition in depth surfaces [28] and planes [27]. Our focus is on the Layered Depth Images (LDI) introduced by Shade et al [35], which refer to a single view representation of a scene that contains multiple layers of RGB-D information.…”
Section: Related Workmentioning
confidence: 99%
“…The advantage is that the 3D mesh captures all the available information in the scene, while an image-based approach [6] only captures information which is present in the set of consecutive image frames to be warped. For every frame, we render the visible instances separately, similarly to [7]. In addition to the color images and the visibility masks, we extract depth maps and object categories for every instance.…”
Section: Data Generationmentioning
confidence: 99%
“…This work is closely related to amodal segmentation works [12], [13], [14]. Generally, amodal perception refers to the intrinsic ability of humans to perceive objects as complete even if they are only partially visible.…”
Section: Introductionmentioning
confidence: 94%
“…For instance, Radford A generated faces in various characteristics. Ehsani K completed the invisible section of scenes. There have already been some GAN‐based attempts on segmentation.…”
Section: Introductionmentioning
confidence: 99%