2018
DOI: 10.1016/j.bica.2018.01.002
|View full text |Cite
|
Sign up to set email alerts
|

Emotion recognition from Marathi speech database using adaptive artificial neural network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(10 citation statements)
references
References 31 publications
0
10
0
Order By: Relevance
“…A black dot (•) in a cell means the corresponding database was used in the research mentioned at the bottom of the column. Year 2005 2010 2011 2013 2014 2016 2017 2018 2019 2020 Research HMM, SVM [6] SVM [17] GerDA, RBM [22] LSTM, BLSTM [28] CRF, CRBM [24] SVM, PCA, LPP, TSL [90] DNN, ANN, ELM [23] DCNN, LSTM [29] CNN [21] DCNN [26] LSTM, MTL [33] ANN, PSOF [19] DCNN, DTPM, TSL [25] LSTM, VAE [31] GAN [86] GAN, SVM [88] LSTM, ATTN [94] DCNN, LSTM [30] CNN, VAE, DAE, AAE, AVB [32] DCNN, GAN [89] LDA, TSL, TLSL [91] CNN, BLSTM, ATTN, MTL [95] LSTM, ATTN [83] DNN, Generative [76] DCNN [79] Additionally, Figure 2a shows a comparison between accuracies reported in deep learning methods based on EMO-DB versus IEMOCAP, which we can see there is a clear separation between the accuracies published. Again, one reason could be the fact that EMO-DB has one degree of magnitude fewer number of samples than IEMOCAP, and using it with deep learning methods makes it more prone to overfitting.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…A black dot (•) in a cell means the corresponding database was used in the research mentioned at the bottom of the column. Year 2005 2010 2011 2013 2014 2016 2017 2018 2019 2020 Research HMM, SVM [6] SVM [17] GerDA, RBM [22] LSTM, BLSTM [28] CRF, CRBM [24] SVM, PCA, LPP, TSL [90] DNN, ANN, ELM [23] DCNN, LSTM [29] CNN [21] DCNN [26] LSTM, MTL [33] ANN, PSOF [19] DCNN, DTPM, TSL [25] LSTM, VAE [31] GAN [86] GAN, SVM [88] LSTM, ATTN [94] DCNN, LSTM [30] CNN, VAE, DAE, AAE, AVB [32] DCNN, GAN [89] LDA, TSL, TLSL [91] CNN, BLSTM, ATTN, MTL [95] LSTM, ATTN [83] DNN, Generative [76] DCNN [79] Additionally, Figure 2a shows a comparison between accuracies reported in deep learning methods based on EMO-DB versus IEMOCAP, which we can see there is a clear separation between the accuracies published. Again, one reason could be the fact that EMO-DB has one degree of magnitude fewer number of samples than IEMOCAP, and using it with deep learning methods makes it more prone to overfitting.…”
Section: Discussionmentioning
confidence: 99%
“…Later in 2018, Darekar and Dhande [ 19 ] have introduced a system based on artificial neural networks, the first extract NMF analysis, Pitch Analysis, and Cepstrum features; then, they reduce their dimensionality applying PCA to their feature vectors. They then feed their features to an artificial neural network introduced by Bhatnagar and Gupta 2017 [ 20 ], called NARX Double Layer, which is an ANN with two hidden layers.…”
Section: Emotion Recognition Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Moreover, the application of neural networks to learn abstract features can avoid manual extraction of features and have local feature abstraction and memory functions. With the development of the artificial intelligence technology, scholars applied the neural network scheme through diverse methods such as through sound (Darekar and Dhande, 2018), facial expression (Jaina et al ., 2019), brain signal (Meza-Kubo et al ., 2016), and the vision (Ruwa et al ., 2019) methods, to evaluate the emotional cognition. Currently, different neural network models, including the back-propagation (BP) neural network, Hopfield neural network, adaptive resonance theory (ART) neural network, and the Kohonen neural network, are used for the emotional prediction evaluation.…”
Section: Related Workmentioning
confidence: 99%
“…In order to address the abovementioned problems and improve the emotion recognition performance, many scholars at home and abroad have studied the recognition algorithms of speech emotion information which are based on the feature learning, and proposed a lot of effective recognition models [6,7]. Literature [8] proposes a multi-level acoustic feature fusion algorithm based on multiple kernel learning. And on this basis, the transfer multiple kernel learning algorithm based multi-level feature fusion is proposed, which can improve the recognition robustness of the feature in the training and the test dataset.…”
Section: Introductionmentioning
confidence: 99%