Proceedings of the 1st ACM International Workshop on Human-Centered Multimedia 2006
DOI: 10.1145/1178745.1178762
|View full text |Cite
|
Sign up to set email alerts
|

Toward multimodal fusion of affective cues

Abstract: During face to face communication, it has been suggested that as much as 70% of what people communicate when talking directly with others is through paralanguage involving multiple modalities combined together (e.g. voice tone and volume, body language). In an attempt to render humancomputer interaction more similar to human-human communication and enhance its naturalness, research on sensory acquisition and interpretation of single modalities of human expressions have seen ongoing progress over the last decad… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
23
0

Year Published

2008
2008
2018
2018

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 36 publications
(23 citation statements)
references
References 42 publications
0
23
0
Order By: Relevance
“…Facial Animation Parameters -FAPs are extracted from video data and are used together with low level audio features as input for a HMM to classify the human emotions. The paper of (Paleari and Lisetti, 2006) presents a multimodal fusion framework for emotion recognition that relies on MAUI -Multimodal Affective User Interface paradigm. The approach is based on the Scherer's theory Component Process Theory (CPT) for the definition of the user model and to simulate the agent emotion generation.…”
Section: Related Workmentioning
confidence: 99%
“…Facial Animation Parameters -FAPs are extracted from video data and are used together with low level audio features as input for a HMM to classify the human emotions. The paper of (Paleari and Lisetti, 2006) presents a multimodal fusion framework for emotion recognition that relies on MAUI -Multimodal Affective User Interface paradigm. The approach is based on the Scherer's theory Component Process Theory (CPT) for the definition of the user model and to simulate the agent emotion generation.…”
Section: Related Workmentioning
confidence: 99%
“…These issues need to be addressed in follow-up studies to obtain a better understanding of the interaction between various expressive cues, sources and modalities in HHI. The multimodal affect systems should potentially be able to detect incongruent messages and label them as incongruent for further/detailed understanding of the information being conveyed (Paleari & Lisetti, 2006) . Different to the cross-mode compensation but still part of the multicue or multimodal perception, there exist findings reporting that when distance is involved humans tend to process the overall global information rather than considering configurations of local regions.…”
Section: Multimodal Expression and Perception Of Emotionsmentioning
confidence: 99%
“…Video features and fNIRS features can be fused at the feature or decision level on a block-by-block basis. (Paleari & Lisetti, 2006) introduce a generic framework with 'resynchronization buffers'. They aim to compare the different estimations, and realign the different evaluations so that they correspond to the same phenomenon even if one estimation is delayed compared to the other one.…”
Section: Challengesmentioning
confidence: 99%
See 1 more Smart Citation
“…In human-interaction, 55% of affective information is carried by the body whilst 38% by the voice tone and volume, and only 7% person by the words spoken [1]. Ekman [2] further suggests that non-verbal behaviours are the primary vehicles for expressing emotion.…”
Section: Introductionmentioning
confidence: 99%