2007
DOI: 10.1016/j.specom.2006.11.004
|View full text |Cite
|
Sign up to set email alerts
|

Ensemble methods for spoken emotion recognition in call-centres

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
166
0
2

Year Published

2010
2010
2015
2015

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 290 publications
(173 citation statements)
references
References 25 publications
5
166
0
2
Order By: Relevance
“…Acoustic features of speech have been used extensively to separate emotional coloring present in the speech signal by employing several pattern recognition techniques (Ang et al, 2002;Nwe et al, 2003;Lee and Narayanan, 2005;Batliner et al, 2006;Schuller et al, 2007b;Kapoor et al, 2007;Morrison et al, 2007;Neiberg and Elenius, 2008;Lee et al, 2009). Phoneme, syllable and word level statistics corresponding to F0 (fundamental frequency), energy, duration, spectral parameters, and voice quality parameters are among the features that have been mainly used for emotion recognition.…”
Section: Introductionmentioning
confidence: 99%
“…Acoustic features of speech have been used extensively to separate emotional coloring present in the speech signal by employing several pattern recognition techniques (Ang et al, 2002;Nwe et al, 2003;Lee and Narayanan, 2005;Batliner et al, 2006;Schuller et al, 2007b;Kapoor et al, 2007;Morrison et al, 2007;Neiberg and Elenius, 2008;Lee et al, 2009). Phoneme, syllable and word level statistics corresponding to F0 (fundamental frequency), energy, duration, spectral parameters, and voice quality parameters are among the features that have been mainly used for emotion recognition.…”
Section: Introductionmentioning
confidence: 99%
“…In the previous studies, most emotion recognition schemes use a single classifier, and very few have considered hybrid classification methods (Morrison, 2007). Intuitively, if the individual schemes can be suitably combined, an improvement in accuracy can be expected.…”
Section: A Hybrid Classification Methodsmentioning
confidence: 99%
“…Scherer (2000) explored the existence of a universal psychobiological mechanism of emotions in speech by studying the recognition of fear, joy, sadness, anger and disgust in nine languages, obtaining 66% of overall accuracy. Two hybrid classification schemes, stacked generalization and the unweighted vote, were proposed and accuracies of 72.18% and 70.54% were achieved respectively, when they were used to recognize anger, disgust, fear, happiness, sadness and surprise (Morrison, 2007). Hybrid classification methods that combined the Support Vector Machines and the Decision Tree were proposed (Nguyen & Bass, 2005).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…While balanced utterances are useful for controlled scientifi analysis and experiments, they may reduce the validity of the data. For this reason, many other researchers prefer that the distribution of the emotions in the database reflect their real-world frequency [104,163]. In this case, the number of neutral utterances should be the largest in the emotional speech corpus.…”
Section: Modeling the User Emotional Statementioning
confidence: 99%