2014
DOI: 10.1007/s12559-014-9296-6
|View full text |Cite
|
Sign up to set email alerts
|

Investigation of Speaker Group-Dependent Modelling for Recognition of Affective States from Speech

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
11
0
2

Year Published

2014
2014
2019
2019

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 23 publications
(13 citation statements)
references
References 77 publications
0
11
0
2
Order By: Relevance
“…As we can see, there is a huge variety in the number of selected features but also in the type of features [19]. Further, from the literature review in [20], we can state that for several of the observed corpora reasonable results were achieved by various research groups using considerably less than 6552 features. According to [7], we also decided to start with a larger feature collection.…”
Section: A Feature Extractionmentioning
confidence: 91%
“…As we can see, there is a huge variety in the number of selected features but also in the type of features [19]. Further, from the literature review in [20], we can state that for several of the observed corpora reasonable results were achieved by various research groups using considerably less than 6552 features. According to [7], we also decided to start with a larger feature collection.…”
Section: A Feature Extractionmentioning
confidence: 91%
“…Through experiments with different datasets it was shown that the recognition performance is significantly improving when considering gender or age. In some cases a combination of both factors achieved an even further improvement [4]. Subsequently, the gender and age-group specific modelling was extended to the fusion of continuous, fragmentary, audio-visual data.…”
Section: Investigated Open Issuesmentioning
confidence: 99%
“…Words which are strongly emphasized by the speaker and hence are perceived as very prominent by the listener frequently indicate a correction after a misunderstanding [41]. Endowing a machine with the capabilities to use this information is the target of our and previous research [26,27]. These words are also visually prominent, i. e. when only observing and not acoutically listening to the speaker [2,5,16,29,40].…”
Section: Introductionmentioning
confidence: 99%
“…These words are also visually prominent, i. e. when only observing and not acoutically listening to the speaker [2,5,16,29,40]. Previously algorithms have been developed to detect these prominent words from the acoustic signal [26,27,33]. In [17] we presented an approach which also integrates information on the speaker's head and mouth movement to detect prominent words.…”
Section: Introductionmentioning
confidence: 99%