2022
DOI: 10.1007/978-3-031-19818-2_41
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Segmentation in Real-World Images via Spelke Object Inference

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
16
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 14 publications
(16 citation statements)
references
References 26 publications
0
16
0
Order By: Relevance
“…It will also be important to use additional means of model evaluation, such as model-matched stimuli 35, 61, 75 , stimuli optimized for the model’s predicted response 76, 88, 89 , or directly substituting brain responses into models 90 . And ultimately, analyses such as these need to be related to more fine-grained anatomy and brain response measurements.…”
Section: Discussionmentioning
confidence: 99%
“…It will also be important to use additional means of model evaluation, such as model-matched stimuli 35, 61, 75 , stimuli optimized for the model’s predicted response 76, 88, 89 , or directly substituting brain responses into models 90 . And ultimately, analyses such as these need to be related to more fine-grained anatomy and brain response measurements.…”
Section: Discussionmentioning
confidence: 99%
“…However, masks recovered from naive image subtraction are imperfect, resulting in worse performance at higher IOU thresholds, which is insufficient for high-performing robotic applications. Optical flow methods can also infer contiguous groups of pixels that move together as the robot is pushing them around [15], [7], [16]. This may be an impractical approach if continuous high-bandwidth video is not available from the robot camera, or if the robot is expected to be grasping objects instead of pushing them in a production environment.…”
Section: B Self-supervised Segmentationmentioning
confidence: 99%
“…An object can be defined as a contiguous group of pixels that move together [7]. When the robot successfully grasps an object, the pixels of that object are removed from the scene.…”
Section: Introductionmentioning
confidence: 99%
“…There's been an explosion of interest in 3D image reconstruction ('A New Trick Lets Artificial Intelligence See in 3D', Wired Magazine [1]), considerable successes in using 3D vision to uncover new biological advances (with DeepMind's AlphaFold [2,3] solving the protein-folding problem), and the suggestion that grounding artificial intelligence in 3D vision will enable better AI (MURI 1 ,(4)(5)(6)). But 3D vision still remains a challenge for AI [7], and is often regarded as the most difficult question facing robotics and autonomous vehicles [8][9][10].At the same time, we are also seeing considerable advances in our understanding of biological vision and navigation. Single-cell recording in freely moving animals has enabled us to understand for the first time how the brain's map of 3D space is organized [11,12], while the emergence of virtual and augmented reality has…”
mentioning
confidence: 99%
“…There's been an explosion of interest in 3D image reconstruction (‘A New Trick Lets Artificial Intelligence See in 3D’, Wired Magazine [1]), considerable successes in using 3D vision to uncover new biological advances (with DeepMind's AlphaFold [2,3] solving the protein-folding problem), and the suggestion that grounding artificial intelligence in 3D vision will enable better AI (MURI 1 , [4–6]). But 3D vision still remains a challenge for AI [7], and is often regarded as the most difficult question facing robotics and autonomous vehicles [8–10].…”
mentioning
confidence: 99%