Machine Audition
DOI: 10.4018/978-1-61520-919-4.ch017
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Emotion Recognition

Abstract: Recent advances in human-computer interaction technology go beyond the successful transfer of data between human and machine by seeking to improve the naturalness and friendliness of user interactions. An important augmentation, and potential source of feedback, comes from recognizing the user‘s expressed emotion or affect. This chapter presents an overview of research efforts to classify emotion using different modalities: audio, visual and audio-visual combined. Theories of emotion provide a framework for de… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 100 publications
(34 citation statements)
references
References 69 publications
(65 reference statements)
0
34
0
Order By: Relevance
“…The second corpus is the SAVEE database, created by Haq and Jackson (2010). This corpus contains speech recordings from four male native English speakers.…”
Section: Experimental Methodologymentioning
confidence: 99%
“…The second corpus is the SAVEE database, created by Haq and Jackson (2010). This corpus contains speech recordings from four male native English speakers.…”
Section: Experimental Methodologymentioning
confidence: 99%
“…We now describe how we employ the CycleGAN model to learn the expression mapping in the blendshape weights space. From the reconstructed blendshape weights of the used training video data set [HJ10], we first sample training expression pairs (xi,yi) independently from a source domain and a target domain. Next, given samples in two expression domains X and Y (e.g.…”
Section: Cycle‐consistent Expression Mappingmentioning
confidence: 99%
“…We used the Surrey Audio‐Visual Expressed Emotion (SAVEE) data set [HJ10] for model training. The data set contains video clips recorded from four male actors with multiple expressions, uttering 120 sentences in English.…”
Section: Cycle‐consistent Expression Mappingmentioning
confidence: 99%
“…The Surrey Audio-Visual Expressed Emotion (SAVEE) database [12] consists of footage of 4 British male actors with six basic emotions(disgust, anger, happy, sad, fear surprise) and neutral state. A total of 480 phonetically balanced sentences are selected from the standard TIMIT corpus [13] for every emotional state.…”
Section: A Datasetsmentioning
confidence: 99%