Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-1349
|View full text |Cite
|
Sign up to set email alerts
|

An Open Source Emotional Speech Corpus for Human Robot Interaction Applications

Abstract: For further understanding the wide array of emotions embedded in human speech, we are introducing a strictly-guided simulated emotional speech corpus. In contrast to existing speech corpora, this was constructed by maintaining an equal distribution of 4 long vowels in New Zealand English. This balance is to facilitate emotion related formant and glottal source feature comparison studies. Also, the corpus has 5 secondary emotions and 5 primary emotions. Secondary emotions are important in Human-Robot Interactio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 24 publications
(21 citation statements)
references
References 20 publications
0
21
0
Order By: Relevance
“…20 subjects participated in this evaluation task. James et al developed an open-source speech corpus [19] for 5 primary emotions and 5 secondary emotions consisting of total 2400 sentences for the New Zealand English. There were 120 participants for the subject evaluation of the corpus.…”
Section: Related Workmentioning
confidence: 99%
“…20 subjects participated in this evaluation task. James et al developed an open-source speech corpus [19] for 5 primary emotions and 5 secondary emotions consisting of total 2400 sentences for the New Zealand English. There were 120 participants for the subject evaluation of the corpus.…”
Section: Related Workmentioning
confidence: 99%
“…The Surrey Audio-Visual Expressed Emotion (SAVEE) [141] database covers 14 speakers with 7 different emotions but only contains 480 utterances in total. JL-Corpus [142] is a collection of 4 New-Zealand English speakers uttering 15 sentences in 5 primary emotions with 2 repetitions, and another 10 sentences in 5 secondary emotions. In general, these databases are all of good quality and can be ideal to build a rule-based emotional voice conversion framework.…”
Section: Lexical Variabilitymentioning
confidence: 99%
“…The ability of the regressor to differentiate between emotions resp. place the emotions in the AV space was tested on ten publicly available databases: EmoDB [19], EMOVO [20], RAVDESS [21], CREMA-D [22], SAVEE [23], VESUS [24], eNTERFACE [25], JL Corpus [26], TESS [27], and GEES [28]. These databases are categorically annotated and do not include information on AV values.…”
Section: Testing Databasesmentioning
confidence: 99%