2014
DOI: 10.1007/978-3-319-10599-4_52
|View full text |Cite
|
Sign up to set email alerts
|

Domain-Adaptive Discriminative One-Shot Learning of Gestures

Abstract: Abstract. The objective of this paper is to recognize gestures in videos -both localizing the gesture and classifying it into one of multiple classes. We show that the performance of a gesture classifier learnt from a single (strongly supervised) training example can be boosted significantly using a 'reservoir' of weakly supervised gesture examples (and that the performance exceeds learning from the one-shot example or reservoir alone). The one-shot example and weakly supervised reservoir are from different 'd… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
56
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 69 publications
(56 citation statements)
references
References 25 publications
0
56
0
Order By: Relevance
“…We follow the experimental protocol adopted in [3,14,22,25] and provide precision, recall and F1-score measures on the validation set. We compare our model with Yao et al [25], Wu et al [22], Pfister et al [14], and Fernando et al [3].…”
Section: Methodsmentioning
confidence: 99%
“…We follow the experimental protocol adopted in [3,14,22,25] and provide precision, recall and F1-score measures on the validation set. We compare our model with Yao et al [25], Wu et al [22], Pfister et al [14], and Fernando et al [3].…”
Section: Methodsmentioning
confidence: 99%
“…Precision Recall F-score Pfister et al [17] 61.2% 62.3% 61.7% Yao et al [28] --56.0% Wu et al [26] 59.9% 59.3% 59.6% VideoDarwin [5] 74.0% 73.8% 73.9% HiVideoDarwin 74.9% 75.6% 74.6% Table 3. Statistical analysis for parameters.…”
Section: Approachmentioning
confidence: 95%
“…For each frame we estimate the body joints using [19] to preprocess these data and extract frame descriptors in the same way as [5]. We report precision, recall, F1-score and mAP on the validation set, as done in [17,28].…”
Section: Datasetsmentioning
confidence: 99%
“…Subtitles have been exploited for assisting the learning of visual recognizer. Several studies [6,8,31] automatically learn British Sign Language signs from TV broadcasts. Their videos contain a single signer with a stable pose.…”
Section: Related Workmentioning
confidence: 99%