2014
DOI: 10.1016/j.procs.2014.11.020
|View full text |Cite
|
Sign up to set email alerts
|

Speech Emotion Recognition in Acted and Spontaneous Context

Abstract: Little attention has been paid so far in the context in which databases used for the study of emotion through vocal channel are recorded. Thus, we propose and evaluate an emotion classification system focusing on the differences between acted and spontaneous emotional speech through the use of two different databases: SAVEE and IEMOCAP. For the purpose of this work, we have examined wavelet packet energy and entropy features applied to Mel, Bark and ERB scale applied with Hidden Markov Model (HMM) as classific… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(3 citation statements)
references
References 14 publications
0
3
0
Order By: Relevance
“…As mentioned, the number of emotional corpora with nonverbal expressions is limited. Emotional speech corpora can be roughly categorized into two types: acted and spontaneous [8]. In acted corpora, scripts and a recording instruction are provided to the speakers before the recording, which makes it easy to control label balance and almost does not need extra manual annotation.…”
Section: B: Emotional Speech Corpus With Nonverbal Expressionsmentioning
confidence: 99%
See 1 more Smart Citation
“…As mentioned, the number of emotional corpora with nonverbal expressions is limited. Emotional speech corpora can be roughly categorized into two types: acted and spontaneous [8]. In acted corpora, scripts and a recording instruction are provided to the speakers before the recording, which makes it easy to control label balance and almost does not need extra manual annotation.…”
Section: B: Emotional Speech Corpus With Nonverbal Expressionsmentioning
confidence: 99%
“…To show the contributions of NVs on emotion recognizability, we also evaluate the verbal parts of JVNV by removing nonverbal parts from the utterances of JVNV, which is denoted as ''JVNV-V''. We conducted a forced choice task on a Japanese crowdsourcing platform 8 . For each corpus, we randomly picked up 60 emotion-balanced samples.…”
Section: B Emotion Recognizabilitymentioning
confidence: 99%
“…Due to these limitations, audio samples in many speechbased emotion datasets are collected through acted elicitation methods relying on individuals who engender a target emotion while uttering pre-determined linguistic contents, also known as scripted speech [6]. Despite the fact that these methods tend to overlook subtle expression details, they provide ample data, based on which machine learning (ML) methodologies can recognize emotions [7].…”
Section: Introductionmentioning
confidence: 99%