2019 International Joint Conference on Neural Networks (IJCNN) 2019
DOI: 10.1109/ijcnn.2019.8851988
|View full text |Cite
|
Sign up to set email alerts
|

Music Artist Classification with Convolutional Recurrent Neural Networks

Abstract: Previous attempts at music artist classification use frame level audio features which summarize frequency content within short intervals of time. Comparatively, more recent music information retrieval tasks take advantage of temporal structure in audio spectrograms using deep convolutional and recurrent models. This paper revisits artist classification with this new framework and empirically explores the impacts of incorporating temporal structure in the feature representation. To this end, an established clas… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
53
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 44 publications
(56 citation statements)
references
References 25 publications
0
53
0
1
Order By: Relevance
“…This result shows that using melody contour as additional features helps SID. Our 'CRNNM+Data aug' model achieves 0.75 song-level F1 score, which is greatly higher than that (0.67) obtained by the best existing model ('CRNN+Origin') [17] for artist20. Table 1 also shows that, for all the three models, training on Data aug outperforms those trained on Origin for the song-level result, validating the effectiveness of the data augmentation method.…”
Section: Resultsmentioning
confidence: 54%
See 4 more Smart Citations
“…This result shows that using melody contour as additional features helps SID. Our 'CRNNM+Data aug' model achieves 0.75 song-level F1 score, which is greatly higher than that (0.67) obtained by the best existing model ('CRNN+Origin') [17] for artist20. Table 1 also shows that, for all the three models, training on Data aug outperforms those trained on Origin for the song-level result, validating the effectiveness of the data augmentation method.…”
Section: Resultsmentioning
confidence: 54%
“…Specifically, we use open-unmix [12], an open-source three-layer bidirectional deep recurrent neural network for SS. Moreover, we build upon our SID model based on the implementation of a convolutional recurrent neural network made available by Nasrullah and Zhao [17], which attains the highest song-level F1-score of 0.67 on the per-album split of the artist20 dataset [18], a standard dataset for SID. As neural networks may find their own way extracting relevant features or patterns from the input, it remains to be studied whether the use of SS can improve the performance of a deep learning based SID model.…”
Section: Conv Blockmentioning
confidence: 99%
See 3 more Smart Citations