2016
DOI: 10.14569/ijacsa.2016.070904
|View full text |Cite
|
Sign up to set email alerts
|

Designing and Implementing of Intelligent Emotional Speech Recognition with Wavelet and Neural Network

Abstract: Abstract-Recognition of emotion from speech is a significant subject in man-machine fields. In this study, speech signal has analyzed in order to create a recognition system which is able to recognize human emotion and a new set of characteristic has proposed in time, frequency and time-frequency domain in order to increase the accuracy. After extracting features of Pitch, MFCC, Wavelet, ZCR and Energy, neural networks classify four emotions of EMO-DB and SAVEE databases. Combination of features for two emotio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 8 publications
0
1
0
Order By: Relevance
“…Additionally, Linear Prediction Coefficients (LPC) and Linear Spectral frequencies (LSF) were used for different applications like speaker recognition [15], spoken digits recognition [16], and emotion recognition from speech [17]. Discrete Wavelet Transform is another feature extraction technique that has been used for speaker recognition [18], speech semantic and emotions recognitions [19,20]. Furthermore, spectrogram images are the best choice of speech and acoustic feature extraction that is suitable for CNN models.…”
Section: Introductionmentioning
confidence: 99%
“…Additionally, Linear Prediction Coefficients (LPC) and Linear Spectral frequencies (LSF) were used for different applications like speaker recognition [15], spoken digits recognition [16], and emotion recognition from speech [17]. Discrete Wavelet Transform is another feature extraction technique that has been used for speaker recognition [18], speech semantic and emotions recognitions [19,20]. Furthermore, spectrogram images are the best choice of speech and acoustic feature extraction that is suitable for CNN models.…”
Section: Introductionmentioning
confidence: 99%