Proceedings of the 2021 International Conference on Multimodal Interaction 2021
DOI: 10.1145/3462244.3479957
|View full text |Cite
|
Sign up to set email alerts
|

HEMVIP: Human Evaluation of Multiple Videos in Parallel

Abstract: In many research areas, for example motion and gesture generation, objective measures alone do not provide an accurate impression of key stimulus traits such as perceived quality or appropriateness. The gold standard is instead to evaluate these aspects through user studies, especially subjective evaluations of video stimuli. Common evaluation paradigms either present individual stimuli to be scored on Likert-type scales, or ask users to compare and rate videos in a pairwise fashion. However, the time and reso… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(19 citation statements)
references
References 12 publications
0
19
0
Order By: Relevance
“…We used a MUSHRA‐like (MUltiple Stimuli with Hidden Reference and Anchor) [ITU15] interface based on [JYW*21]. We had a total of 131 participants, with a minimum of 20 per study (ages 20‐55 years μ = 33.6, σ = 8.1).…”
Section: Methodsmentioning
confidence: 99%
“…We used a MUSHRA‐like (MUltiple Stimuli with Hidden Reference and Anchor) [ITU15] interface based on [JYW*21]. We had a total of 131 participants, with a minimum of 20 per study (ages 20‐55 years μ = 33.6, σ = 8.1).…”
Section: Methodsmentioning
confidence: 99%
“…We use a MUSHRA-like [21] (MUltiple Stimuli with Hidden Reference and Anchor) interface commonly used for subjective evaluation of speech-synthesis [44], but here adapted for video interfaces, since such setups have been found to work well for evaluating head motion and hand gestures [7,23,33]. On a single test page, participants are presented with videos of generated gesture-speech from all evaluated models on the same input text sentence.…”
Section: Perceptual Evaluation Methodsmentioning
confidence: 99%
“…One important aspect to evaluate for gesture‐generation systems is the human‐likeness of the generated gestures, which is measured and compared through human perceptual studies, often with comparable stimuli presented side by side as in e.g. [JYW*21, KJY*21, WGKB21]. On the other hand, evaluating the other aspects such as the appropriateness and/or specificity of generated gestures in the context of speech and other multimodal grounding information (see Section 6.4) is quite challenging, especially since differences in the human‐likeness of the motions being compared tends to interfere with perceived gesture appropriateness (cf.…”
Section: Key Challenges Of Gesture Generationmentioning
confidence: 99%