2021
DOI: 10.48550/arxiv.2104.09958
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

GENESIS-V2: Inferring Unordered Object Representations without Iterative Refinement

Abstract: Advances in object-centric generative models (OCGMs) have culminated in the development of a broad range of methods for unsupervised object segmentation and interpretable object-centric scene generation. These methods, however, are limited to simulated and real-world datasets with limited visual complexity. Moreover, object representations are often inferred using RNNs which do not scale well to large images or iterative refinement which avoids imposing an unnatural ordering on objects in an image but requires… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(19 citation statements)
references
References 32 publications
0
19
0
Order By: Relevance
“…Comparison to prior work. In the appendix, we further compare our approach with two recent end-to-end object-centric models (Lin et al, 2020b;Engelcke et al, 2021). Both models struggle to scale to the FISHBOWL dataset, highlighting the advantage of our multi-stage approach compared to end-to-end scene models.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations

Unsupervised Object Learning via Common Fate

Tangemann,
Schneider,
von Kügelgen
et al. 2021
Preprint
“…Comparison to prior work. In the appendix, we further compare our approach with two recent end-to-end object-centric models (Lin et al, 2020b;Engelcke et al, 2021). Both models struggle to scale to the FISHBOWL dataset, highlighting the advantage of our multi-stage approach compared to end-to-end scene models.…”
Section: Methodsmentioning
confidence: 99%
“…Those approaches typically use additional supervision such as ground-truth segmentation or additional views, with Ehrhardt et al (2020); Niemeyer & Geiger (2021) being notable exceptions. While most methods can decompose a given scene into its constituent objects, only few are fully-generative in the sense that they can generate novel scenes (Lin et al, 2020a;Ehrhardt et al, 2020;von Kügelgen et al, 2020;Engelcke et al, 2021;Niemeyer & Geiger, 2021;Dittadi & Winther, 2019).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation

Unsupervised Object Learning via Common Fate

Tangemann,
Schneider,
von Kügelgen
et al. 2021
Preprint
“…There are many parametric forms M could employ. For instance, one could consider generating multiple masks sequentially in an auto-regressive manner as seen in Engelcke et al (2021). However, we find that the simplest setup suffices, where M : R c×w×h → R N ×w×h consists of a learnable neural network and a pixelwise softmax layer σ applied across N masks to ensure that the sum of a given pixel across all the masks equals 1.…”
Section: Occlusion Model Mmentioning
confidence: 99%
“…Datasets We perform a qualitative evaluation using the realworld Sketchy dataset [23], which contains demonstration trajectories of a robot arm performing different tasks involving a set of objects. The images are pre-processed as in [37], using a 128 × 128 resolution and sequences of length 10. Sketchy, however, does not contain object annotations, which prohibits the quantitative evaluation of object segmentation and tracking methods.…”
Section: Learningmentioning
confidence: 99%