2005 IEEE International Conference on Multimedia and Expo
DOI: 10.1109/icme.2005.1521551
|View full text |Cite
|
Sign up to set email alerts
|

Audio-Visual Affect Recognition in Activation-Evaluation Space

Abstract: The ability of a computer to detect and appropriately respond to changes in a user's affective state has significant implications to Human-Computer Interaction (HCI). To more accurately simulate the human ability to assess affects through multi-sensory data, automatic affect recognition should also make use of multimodal data. In this paper, we present our efforts toward audio-visual affect recognition. Based on psychological research, we have chosen affect categories based on an activationevaluation space whi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 13 publications
(5 citation statements)
references
References 9 publications
0
5
0
Order By: Relevance
“…Yu et al (2004) classified user engagement in social telephone conversations between friends along arousal and valence scales that were discretized into 5 levels. Kim et al (2005), Zeng et al (2005) andWo ¨llmer et al (2009) classified emotions in the 4 emotion quadrants of the arousal-valence space. Instead of classifying emotions on discretized scales of arousal and valence, some studies have taken up the challenge to classify emotions on continuous scales of arousal and valence.…”
Section: Related Workmentioning
confidence: 99%
“…Yu et al (2004) classified user engagement in social telephone conversations between friends along arousal and valence scales that were discretized into 5 levels. Kim et al (2005), Zeng et al (2005) andWo ¨llmer et al (2009) classified emotions in the 4 emotion quadrants of the arousal-valence space. Instead of classifying emotions on discretized scales of arousal and valence, some studies have taken up the challenge to classify emotions on continuous scales of arousal and valence.…”
Section: Related Workmentioning
confidence: 99%
“…• Audio-Visual Classification may be useful in identifying affective cues such as vocal inflections or prosodic features [180], [188], [223].…”
Section: Affect Recognitionmentioning
confidence: 99%
“…Among the audiovisual affect recognition approaches, Busso et al [190], Petridis and Pantic [184] and Schuller et al [183] have employed feature-level fusion, which concatenates multimodal features and passes them through a single affect recognition model or classifier. Decision-level fusion, on the other hand, has been used by Hoch et al [255], Go et al [187], Pal et al [185], Wang and Guan [180], Zeng et al [178], Zeng et al [256], and Zeng et al [223]. However, decision-level fusion ignores the inherent correlation between these multi-modal features.…”
Section: G Audio-visual Analysis In Affect Recognitionmentioning
confidence: 99%
See 2 more Smart Citations