2012
DOI: 10.1007/s13735-012-0026-0
|View full text |Cite
|
Sign up to set email alerts
|

Tonal representations for music retrieval: from version identification to query-by-humming

Abstract: In this study we compare the use of different music representations for retrieving alternative performances of the same musical piece, a task commonly referred to as version identification. Given the audio signal of a song, we compute descriptors representing its melody, bass line and harmonic progression using state-of-the-art algorithms. These descriptors are then employed to retrieve different versions of the same musical piece using a dynamic programming algorithm based on nonlinear time series analysis. F… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
38
0
3

Year Published

2014
2014
2021
2021

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 69 publications
(41 citation statements)
references
References 27 publications
0
38
0
3
Order By: Relevance
“…In other scenarios, one may have to deal with imperfect or less specific queries. For instance, a query might be a person humming the solo, which then requires an extra step for extracting the fundamental frequency from the hummed melody (queryby-humming), see, e.g., Pauws (2002), Ryynänen and Klapuri (2008), and Salamon et al (2013). Furthermore, the query may have a different tuning, may be transposed to another key, or played with rhythmic variations.…”
Section: Perspectivesmentioning
confidence: 99%
See 1 more Smart Citation
“…In other scenarios, one may have to deal with imperfect or less specific queries. For instance, a query might be a person humming the solo, which then requires an extra step for extracting the fundamental frequency from the hummed melody (queryby-humming), see, e.g., Pauws (2002), Ryynänen and Klapuri (2008), and Salamon et al (2013). Furthermore, the query may have a different tuning, may be transposed to another key, or played with rhythmic variations.…”
Section: Perspectivesmentioning
confidence: 99%
“…A common preprocessing step, which targets the "polyphony gap" between query and database document, is to enhance the predominant melody in audio recordings. In Salamon et al (2013), the authors used a so-called salience representation in a query-by-humming system which led to a substantial increase in performance (Salamon and Gómez, 2012). In Balke et al (2017b), a data-driven approach is used to estimate a salience representation for jazz music recordings which showed a similar performance as the aforementioned, salience-based method.…”
Section: Perspectivesmentioning
confidence: 99%
“…Evaluations have been conducted using a publicly available dataset [22], [23], released for the purpose of evaluation on singing transcription. The dataset consists of 38 audio recordings of monophonic singing, recorded with a sample rate of 44.1 kHz and a 16-bit resolution.…”
Section: Datasetmentioning
confidence: 99%
“…One of the first proposed systems is MUSART by Birmingham et al [43]. Music collections for this task were traditionally built with music scores, user hummed or tapped queries -more recently with audio signals as in the system by Salamon et al [218]. Commercial systems are also exploiting the idea of retrieving music by singing, humming or typing.…”
Section: Music Retrievalmentioning
confidence: 99%