Proceedings of the 6th International Conference on Multimodal Interfaces 2004
DOI: 10.1145/1027933.1027958
|View full text |Cite
|
Sign up to set email alerts
|

Bimodal HCI-related affect recognition

Abstract: Perhaps the most fundamental application of affective computing would be Human-Computer Interaction (HCI) in which the computer is able to detect and track the user's affective states, and make corresponding feedback. The human multi-sensor affect system defines the expectation of multimodal affect analyzer. In this paper, we present our efforts toward audio-visual HCI-related affect recognition. With HCI applications in mind, we take into account some special affective states which indicate users' cognitive/m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
45
0
2

Year Published

2005
2005
2007
2007

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 52 publications
(48 citation statements)
references
References 12 publications
1
45
0
2
Order By: Relevance
“…They could improve the performance of decision-level fusion by considering the dominant modality, determined by empirical studies, in case significant discrepancy between the outputs of each unimodal classifier has been observed. Recently, a large-scale audio-visual dataset was collected by (Zeng et al, 2004), which contains five HCI-related affective responses (confusion, interest, boredom, and frustration) in addition to seven affects (the six basic emotions + neutral). To classify the 11 emotions subject-dependently, they used the SNoW (Sparse Network of Winnow) classifier with Naive Bayes as the update rule and achieved a recognition accuracy of almost 90% through bimodal fusion while the unimodal classifiers yielded only 45-56%.…”
Section: Automatic Emotion Recognition Using Speech and Biosignalsmentioning
confidence: 99%
“…They could improve the performance of decision-level fusion by considering the dominant modality, determined by empirical studies, in case significant discrepancy between the outputs of each unimodal classifier has been observed. Recently, a large-scale audio-visual dataset was collected by (Zeng et al, 2004), which contains five HCI-related affective responses (confusion, interest, boredom, and frustration) in addition to seven affects (the six basic emotions + neutral). To classify the 11 emotions subject-dependently, they used the SNoW (Sparse Network of Winnow) classifier with Naive Bayes as the update rule and achieved a recognition accuracy of almost 90% through bimodal fusion while the unimodal classifiers yielded only 45-56%.…”
Section: Automatic Emotion Recognition Using Speech and Biosignalsmentioning
confidence: 99%
“…The number of 30 features seems to be good point where the training and testing errors are about 0.14 and 0.16 on evaluation axis, and 0.11 and 0.13 on activation axis. These coarse-emotion-category-based performances are greatly above our previous fineemotion-category-based classification at the frame level [11]. …”
Section: Methodsmentioning
confidence: 57%
“…In addition, [11] considered four HCI-related cognitive states besides six basic emotions. Contrary to the fine categories of emotion representation, this paper explores the coarse categories for automatic affect recognition.…”
Section: Related Workmentioning
confidence: 99%
“…This variant is manifested in very different application areas including among others data input (audio-visual speech recognition), person identification [39], emotion recognition [44].…”
Section: Known Usesmentioning
confidence: 99%