2022
DOI: 10.1155/2022/5732687
|View full text |Cite
|
Sign up to set email alerts
|

Music Emotion Recognition Model Using Gated Recurrent Unit Networks and Multi-Feature Extraction

Abstract: A large number of music platforms have appeared on the Internet recently. The deep learning framework for music recommendation is still very limited when it comes to accurately identifying the emotional type of music and recommending it to users. Languages, musical styles, thematic scenes, and the ages to which they belong are all common classifications. And this is far from sufficient, posing difficulties in music classification and identification. As a result, this paper uses the methods of music emotion mul… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 21 publications
0
3
0
Order By: Relevance
“…The BiGRU emotion recognition model is developed in another research, and it is compared to other models. Up to 79 % and 81.01 % of the time, respectively, BiGRU can properly detect music with joyful and sad emotions [26]. According to the findings, Naive Bayes had the highest classification accuracy for musical mood, at 86.64% [27].…”
Section: B Related Workmentioning
confidence: 92%
See 1 more Smart Citation
“…The BiGRU emotion recognition model is developed in another research, and it is compared to other models. Up to 79 % and 81.01 % of the time, respectively, BiGRU can properly detect music with joyful and sad emotions [26]. According to the findings, Naive Bayes had the highest classification accuracy for musical mood, at 86.64% [27].…”
Section: B Related Workmentioning
confidence: 92%
“…BiGRU model, convolutional long short-term memory deep neural network (CLDNN) model, CNN-LSTM model, etc. are recently proposed deep learning algorithms to classify musical emotions [26] [38] [39]. But previous researches had the flaws of low accuracy and overfitting problem.…”
Section: A Introductionmentioning
confidence: 99%
“…Pooling layers are typically used after convolution layers to perform nonlinear down sampling by taking the maximum or average value for each small subsection of the matrix. Our model comprises two consecutive onedimensional convolution layers that use 32 and 16 lters with ReLU [30] as an activation function, a kernel size of 7, stride 1, and max pooling of size 4 and stride 4. We applied batch normalization [31] after each convolutional layer to stabilize the network.…”
Section: Model Descriptionmentioning
confidence: 99%