2010
DOI: 10.1109/tpami.2009.107
|View full text |Cite
|
Sign up to set email alerts
|

Classification of Complex Information: Inference of Co-Occurring Affective States from Their Expressions in Speech

Abstract: We present a classification algorithm for inferring affective states (emotions, mental states, attitudes, and the like) from their nonverbal expressions in speech. It is based on the observations that affective states can occur simultaneously and different sets of vocal features, such as intonation and speech rate, distinguish between nonverbal expressions of different affective states. The input to the inference system was a large set of vocal features and metrics that were extracted from each utterance. The … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
22
0

Year Published

2010
2010
2015
2015

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 39 publications
(22 citation statements)
references
References 34 publications
0
22
0
Order By: Relevance
“…Emotions are visualized through various indicators in humans, many of these indicators have been previously analyzed to provide affective knowledge to machines, focusing on facial expressions [5], [6], vocal features [7], [8], [9], body movements and postures [10], [11], [12], [13] and the integration of all of them in emotion analysis systems [14], [15], [16]. But human beings cannot always hope that robots may be able to react in a timely and sensible manner, especially if they haven't be able to recover all the affective information through their sensors.…”
Section: Introductionmentioning
confidence: 99%
“…Emotions are visualized through various indicators in humans, many of these indicators have been previously analyzed to provide affective knowledge to machines, focusing on facial expressions [5], [6], vocal features [7], [8], [9], body movements and postures [10], [11], [12], [13] and the integration of all of them in emotion analysis systems [14], [15], [16]. But human beings cannot always hope that robots may be able to react in a timely and sensible manner, especially if they haven't be able to recover all the affective information through their sensors.…”
Section: Introductionmentioning
confidence: 99%
“…Popular classification models used include, among others, different decision trees [25], support vector machines [5], [26], [27], [17], [6], neural networks [12], [10] and Hidden Markov Models. Again, which is the best classifier often depends on the application and corpus.…”
Section: Previous Researchmentioning
confidence: 99%
“…This subset consisted of 548 samples spoken by 10 different actors. Each category contains samples from the groups as [5], [25], allowing direct comparison of results.…”
Section: Emotion Classificationmentioning
confidence: 99%
“…The audio analysis component used in the research is an adaptation of the framework introduced by Sobol-Shikler [17] and enhanced by Pfister [18]. The OpenSMILE library extracts 6555 features which represent pitch, spectral envelope, and energy feature groups (amongst others); delta and acceleration information; and their functionals (e.g.…”
Section: Audio Analysismentioning
confidence: 99%