2013
DOI: 10.1007/978-3-642-38886-6_35
|View full text |Cite
|
Sign up to set email alerts
|

Detecting Hand-Head Occlusions in Sign Language Video

Abstract: A large body of current linguistic research on sign language is based on analyzing large corpora of video recordings. This requires either manual or automatic annotation of the videos. In this paper we introduce methods for automatically detecting and classifying hand-head occlusions in sign language videos. Linguistically, hand-head occlusions are an important and interesting subject of study as the head is a structural place of articulation in many signs. Our method combines easily calculable local video pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 7 publications
0
5
0
Order By: Relevance
“…To the best of our knowledge, few research contributions have proposed evaluation metrics for sign spotting. In [13], a tool to assess the spotting performance on continuous signing was developed. The evaluation is symmetrical, which is not considered ideal for retrieval metrics [24].…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…To the best of our knowledge, few research contributions have proposed evaluation metrics for sign spotting. In [13], a tool to assess the spotting performance on continuous signing was developed. The evaluation is symmetrical, which is not considered ideal for retrieval metrics [24].…”
Section: Discussionmentioning
confidence: 99%
“…Like the sign spotting metric in [13], audio segmentation metrics and tolerances are used in a symmetrical manner. We distinguish our approach from them by proposing an asymmetrical evaluation.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…We fine tune the last hidden layer of the model while keeping the rest of the layers frozen to ensure that all learning of sign language is embedded in the representations of this last layer. After exploration, we determine that the average length of signs in the CNGT is approximately 12 frames, corresponding to previous finds that co-articulated signs have length of approximately 7 to 13 frames [19,20,21]. Thus, we set the number of frames per input video to 16.…”
Section: Methodsmentioning
confidence: 96%
“…We would like to develop a system that can assist any monitoring system in telling if an infant is in an improper sleeping position, causing her face being covered by the bedding. Existing works inferring occlusion on head images, such as [7,8,9], do not focus on analyzing infant images. On the contrary, we introduce a novel dataset named YunInfants and propose a method focusing on the infant group, aiming to handle any head pose and imaging under varying lighting conditions.…”
Section: Introductionmentioning
confidence: 99%