2020
DOI: 10.48550/arxiv.2003.12352
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Enhanced Self-Perception in Mixed Reality: Egocentric Arm Segmentation and Database with Automatic Labelling

Ester Gonzalez-Sosa,
Pablo Perez,
Ruben Tolosana
et al.

Abstract: In this study, we focus on the egocentric segmentation of arms to improve self-perception in Augmented Virtuality (AV). The main contributions of this work are: i) a comprehensive survey of segmentation algorithms for AV; ii) an Egocentric Arm Segmentation Dataset, composed of more than 10, 000 images, comprising variations of skin color, and gender, among others. We provide all details required for the automated generation of groundtruth and semi-synthetic images; iii) the use of deep learning for the first t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
10
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(10 citation statements)
references
References 54 publications
0
10
0
Order By: Relevance
“…Pigny et Dominjon [11] were the first to propose a deep algorithm to segment egocentric bodies. Their architecture was based on U-NET and trained using a hybrid dataset composed of images from the COCO dataset belonging to persons and a 1500-image custom dataset created following the automatic labelling procedure reported in [6]. They reported 16 ms of inference time for 256 × 256 images, and also observed problems of false positives that downgrade the AV experience.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations

Egocentric Human Segmentation for Mixed Reality

Gajic,
Gonzalez-Sosa,
Gonzalez-Morin
et al. 2020
Preprint
Self Cite
“…Pigny et Dominjon [11] were the first to propose a deep algorithm to segment egocentric bodies. Their architecture was based on U-NET and trained using a hybrid dataset composed of images from the COCO dataset belonging to persons and a 1500-image custom dataset created following the automatic labelling procedure reported in [6]. They reported 16 ms of inference time for 256 × 256 images, and also observed problems of false positives that downgrade the AV experience.…”
Section: Related Workmentioning
confidence: 99%
“…They reported 16 ms of inference time for 256 × 256 images, and also observed problems of false positives that downgrade the AV experience. In this work we plan to extend our previous work [6] and explore further this segmentation problem targeting to perform real-time segmentation at higher resolutions, which are needed for creating and achieving a realistic immersive experience.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation

Egocentric Human Segmentation for Mixed Reality

Gajic,
Gonzalez-Sosa,
Gonzalez-Morin
et al. 2020
Preprint
Self Cite
“…The ability to localize human body parts from images is an essential requirement for many robotics and perception applications such as human-robot interaction [14], surgical robotics and medical images analysis [1], self-perception in mixed reality [9], and 3D human body reconstruction [25]. A collaborative robot, for example, should be able to locate the arms of the human operator with whom it is collaborating, just as a surgical robot must be able to accurately locate the limb on which it is operating.…”
Section: Introductionmentioning
confidence: 99%
“…Background removal systems can be used for body part localization as well: unfortunately, the generality of those systems allows to obtain acceptable results only in specific cases where the foreground contains only human limbs, or when a rich prior knowledge about the scene is given [18]. Other techniques, such as background removal in a green screen setting [9], allow to speed up the annotation process. However, they may come at the cost of the unrealistic feature of green color bleeding at the annotation boundaries.…”
Section: Introductionmentioning
confidence: 99%