2022
DOI: 10.1088/1742-6596/2236/1/012003
|View full text |Cite
|
Sign up to set email alerts
|

Speech Emotion Based Sentiment Recognition using Deep Neural Networks

Abstract: The capacity to comprehend and communicate with others via language is one of the most valuable human abilities. We are well-trained in our experience reading awareness of different emotions since they play a vital part in communication. Contrary to popular belief, emotion recognition is a challenging task for computers or robots due to the subjective nature of human mood. This research proposes a framework for acknowledging the passionate sections of conversation, independent of the semantic content, via the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(5 citation statements)
references
References 14 publications
0
5
0
Order By: Relevance
“…Gradient Boosting excels with 84.96% accuracy on the merged dataset. The datasets RAVDESS and TESS datasets were integrated using CNN, yielding a 97.1% accuracy in [42]. RAVDESS, TESS and SAVEE datasets were integrated using neural network yielding a testing accuracy of about 89.26% in [43].…”
Section: Resultsmentioning
confidence: 99%
“…Gradient Boosting excels with 84.96% accuracy on the merged dataset. The datasets RAVDESS and TESS datasets were integrated using CNN, yielding a 97.1% accuracy in [42]. RAVDESS, TESS and SAVEE datasets were integrated using neural network yielding a testing accuracy of about 89.26% in [43].…”
Section: Resultsmentioning
confidence: 99%
“…5 shows the Random Forest architecture. Double randomness is the random forest's primary attribute [17]. Random Forest is renowned for its resistance to overfitting and capacity for handling large-scale, multidimensional data.…”
Section: Classification Methodsmentioning
confidence: 99%
“…MFCC features are based on human hearing perception [14]. The first step in the MFCC feature extraction procedure is to split the recorded speech data into brief frames, which typically last 20-40 milliseconds and have a minimal amount of overlap.…”
Section: B Extraction Of Featuresmentioning
confidence: 99%
“…Koduru et al [19] investigated MFCCs, LPCs, and spectrograms with SVMs, achieving 81.25% on RAVDESS. Choudhary et al [20] proposed a CNN-RNN system with MFCCs, chroma, and tonnetz, reporting 87.5% on RAVDESS. Dutt and Gader [21] used wavelet decomposition with 1D CNN-LSTM, achieving 84.6% on RAVDESS.…”
Section: Literature Reviewmentioning
confidence: 99%