2023
DOI: 10.1016/j.apacoust.2023.109425
|View full text |Cite
|
Sign up to set email alerts
|

Generative emotional AI for speech emotion recognition: The case for synthetic emotional speech augmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(2 citation statements)
references
References 24 publications
0
2
0
Order By: Relevance
“…Özseven [54] discussed the effectiveness of the spectrogram as a time-frequency domain image in urban sound classification. Latif et al [55], Shafik et al [56] and Mushtaq et al [57] used the spectrogram as an effective acoustic feature in speech emotion recognition, speaker identification, and environmental sound classification, respectively. All of these approaches were based on deep learning.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Özseven [54] discussed the effectiveness of the spectrogram as a time-frequency domain image in urban sound classification. Latif et al [55], Shafik et al [56] and Mushtaq et al [57] used the spectrogram as an effective acoustic feature in speech emotion recognition, speaker identification, and environmental sound classification, respectively. All of these approaches were based on deep learning.…”
Section: Literature Reviewmentioning
confidence: 99%
“…In [7] The paper proposes an end-to-end multi-speaker emotional text-to-speech system with a condition encoder to modulate speaker voice features and emotions in the output. It then utilizes the synthesized emotional speech to augment speech emotion recognition systems, showing improvements in performance.…”
Section: Literature Surveymentioning
confidence: 99%