2019
DOI: 10.1016/j.patrec.2019.08.021
|View full text |Cite
|
Sign up to set email alerts
|

Handwritten Music Recognition for Mensural notation with convolutional recurrent neural networks

Abstract: Optical Music Recognition is the technology that allows computers to read music notation, which is also referred to as Handwritten Music Recognition when it is applied over handwritten notation. This technology aims at efficiently transcribing written music into a representation that can be further processed by a computer. This is of special interest to transcribe the large amount of music written in early notations, such as the Mensural notation, since they represent largely unexplored heritage for the musico… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
44
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 53 publications
(44 citation statements)
references
References 16 publications
0
44
0
Order By: Relevance
“…In the previous work on which our approach is based [6,7], the network outputs these sequences with an agnostic (graphical) codification, which does not follow any standard digital music representation, as it only represents the music symbols by physical features such as the shape and its position in the staff. However, we will train the network to output a semantic encoding directly.…”
Section: Neural Approachmentioning
confidence: 99%
See 2 more Smart Citations
“…In the previous work on which our approach is based [6,7], the network outputs these sequences with an agnostic (graphical) codification, which does not follow any standard digital music representation, as it only represents the music symbols by physical features such as the shape and its position in the staff. However, we will train the network to output a semantic encoding directly.…”
Section: Neural Approachmentioning
confidence: 99%
“…There are other approaches which do not have this limitation, such as the seq2seq approach. However, they have proven to be much less effective than the CRNN/CTC model for OMR [7]. What is important here is that the aforementioned operation of the CRNN/CTC approach forces us to directly discard highly verbose encodings, such as MEI [16] or MusicXML [13], as output formats to be considered.…”
Section: Output Codificationmentioning
confidence: 99%
See 1 more Smart Citation
“…Further testing on the identical data sets using additional metrics is necessary before we might reach firmer conclusions in this respect. Since other studies involve a set of additional criteria that we have not investigated to date, it is also difficult to compare our results with recent systems that focus on early handwritten music notations [16,38]; again this is a potential area for future investigation with IntelliOMR.…”
Section: Future Workmentioning
confidence: 99%
“…Arising out of the HISPAMUS project, Rizo et al [50] describe an effective online tool, MUsic Recognition, Encoding, and Transcription (MuRET ), for managing the OMR and encoding of monophonic sources, including techniques signalled in [38]. Lastly, Calvo-Zaragoza et al [16] utilize a hybrid approach, using traditional algorithms for deskewing the image and identifying staves and normalising the visual area to process, followed by a recurrent convolutional neural network, resulting in "effective transcription" of handwritten mensural notation of 93% accuracy. Also notable is recent work that uses full CNNs on early square music notation to achieve a 96% accuracy in symbol identification [56].…”
Section: Introductionmentioning
confidence: 99%