2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018
DOI: 10.1109/icassp.2018.8461807
|View full text |Cite
|
Sign up to set email alerts
|

Extended Pipeline for Content-Based Feature Engineering in Music Genre Recognition

Abstract: We present a feature engineering pipeline for the construction of musical signal characteristics, to be used for the design of a supervised model for musical genre identification. The key idea is to extend the traditional two-step process of extraction and classification with additive stand-alone phases which are no longer organized in a waterfall scheme. The whole system is realized by traversing backtrack arrows and cycles between various stages. In order to give a compact and effective representation of the… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…As shown in The SOTA methods GTZAN dataset (%) Bisharad et al [7] 85.36 Bisharad et al [8] 82.00 Raissi et al [42] 91.00 Sugianto et al [45] 71.87 Ashraf et al [3] 87.79 Ng et al [39] (FusionNet) 96.50 Liu et al [30] 93.90 Nanni et al [37] 90.60 Ours (MS-SincResNet) 91.49…”
Section: Ablation Studymentioning
confidence: 99%
“…As shown in The SOTA methods GTZAN dataset (%) Bisharad et al [7] 85.36 Bisharad et al [8] 82.00 Raissi et al [42] 91.00 Sugianto et al [45] 71.87 Ashraf et al [3] 87.79 Ng et al [39] (FusionNet) 96.50 Liu et al [30] 93.90 Nanni et al [37] 90.60 Ours (MS-SincResNet) 91.49…”
Section: Ablation Studymentioning
confidence: 99%
“…The research presented in this section have a common goal of condensing the dataset by striping out non-essential data points while maintaining the critical information for the NN to learn from. The features extracted from the dataset should be comprehensive, compact, and effective; this means they should represent the music well, require a smaller amount of storage, and require little computation to extract [1,[48][49][50]. If the correct features are not extracted or there is any loss of data while extracting, the machine learning phase will lack the ability to make use of vital information and thus the features chosen will significantly affect the final results and accuracy of the work [40].…”
Section: Music Genre Classification Focused On Engineered Featuresmentioning
confidence: 99%
“…LSTM) can grasp the prominent long-term dependency based properties, such as recurrent harmonics and music structure contained in the music. These are the possible reasons why deep learning architecture based schemes have achieved tremendous success in various MIR tasks, such as onset detection [6], emotion recognition [7], chord estimation [8], rhythm stimuli recognition [9], source separation [10], music recommendation [11] and auto-tagging [4], [12], [14], [15]. For music classification tasks, CNN and RNN are the two most adopted deep learning architectures.…”
Section: Introductionmentioning
confidence: 99%