2012 19th IEEE International Conference on Image Processing 2012
DOI: 10.1109/icip.2012.6467133
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised classification of extreme facial events using active appearance models tracking for sign language videos

Abstract: We propose an Unsupervised method for Extreme States Classification (UnESC) on feature spaces of facial cues of interest. The method is built upon Active Appearance Models (AAM) face tracking and on feature extraction of Global and Local AAMs. UnESC is applied primarily on facial pose, but is shown to be extendable for the case of local models on the eyes and mouth. Given the importance of facial events in Sign Languages we apply the UnESC on videos from two sign language corpora, both American (ASL) and Greek… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
3
2
1

Relationship

3
3

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 12 publications
0
9
0
Order By: Relevance
“…Overall, different methods focus partially on some of the above aspects, and to the best of our knowledge, none of them shares all described issues. In [43], we introduced ESC. Herein, the approach is extensively presented with mature and updated material, including the application to multiple events and more experiments.…”
Section: Relative Literaturementioning
confidence: 99%
See 1 more Smart Citation
“…Overall, different methods focus partially on some of the above aspects, and to the best of our knowledge, none of them shares all described issues. In [43], we introduced ESC. Herein, the approach is extensively presented with mature and updated material, including the application to multiple events and more experiments.…”
Section: Relative Literaturementioning
confidence: 99%
“…This spreading is controlled through a parameter that we call Subjective Perceived Threshold (SPThres). In our previous work [43], we formed clusters by selecting pairs of data points that have maximum distance until reaching each time the selected scalar SPThres value. In this work, the SPThres value is determined depending on the introduced facial event symmetry.…”
Section: Cluster Selectionmentioning
confidence: 99%
“…Analogous to phonemes in speech, sign languages can be broken down into cheremes, the smallest distinctive structural units [58]. Cheremes can be represented as motion primitives, a set of manual and non-manual motions 1 that are combined to represent all sign language utterances. Such phonetic representations are typically used by linguists for annotation [29,62] or in graphical based avatars for sign generation.…”
Section: Introductionmentioning
confidence: 99%
“…They also discover a complex interaction between head position and mouth shape. Antonakos et al [2], [3] propose a novel semi-supervised approach for Extreme States Classification (ESC) on feature spaces of facial cues in SL videos. Their method is built upon AAM face tracking and feature extraction of global and local AAMs and applied for detection of sign boundaries and alternative constructions.…”
Section: Mouth Non-manuals In Existing Aslr Systemsmentioning
confidence: 99%