2022 25th International Conference on Information Fusion (FUSION) 2022
DOI: 10.23919/fusion49751.2022.9841342
|View full text |Cite
|
Sign up to set email alerts
|

Deep Learning for Audio Visual Emotion Recognition

Abstract: Human emotions can be presented in data with multiple modalities, e.g. video, audio and text. An automated system for emotion recognition needs to consider a number of challenging issues, including feature extraction, and dealing with variations and noise in data. Deep learning have been extensively used recently, offering excellent performance in emotion recognition. This work presents a new method based on audio and visual modalities, where visual cues facilitate the detection of the speech or non-speech fra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 46 publications
0
1
0
Order By: Relevance
“…Different studies have applied neural network-based analyses with speech data to address a range of goals, such as automatic speech recognition (Graves and Jaitly, 2014;Wu et al, 2020), automatic speaker recognition (Devi et al, 2020), speech emotion recognition based on audio (Jiang et al, 2019) and audiovisual information (Hussain et al, 2022), speech enhancement for cochlear implants-aiming to improve the clarity and quality of speech comprehension for individuals with hearing loss (Kang et al, 2021), and, the assessment of speech/voice impairments in chronic degenerative disorders (Maskeliūnas et al, 2022).…”
Section: Introductionmentioning
confidence: 99%
“…Different studies have applied neural network-based analyses with speech data to address a range of goals, such as automatic speech recognition (Graves and Jaitly, 2014;Wu et al, 2020), automatic speaker recognition (Devi et al, 2020), speech emotion recognition based on audio (Jiang et al, 2019) and audiovisual information (Hussain et al, 2022), speech enhancement for cochlear implants-aiming to improve the clarity and quality of speech comprehension for individuals with hearing loss (Kang et al, 2021), and, the assessment of speech/voice impairments in chronic degenerative disorders (Maskeliūnas et al, 2022).…”
Section: Introductionmentioning
confidence: 99%