Proceedings of the 28th ACM International Conference on Multimedia 2020
DOI: 10.1145/3394171.3413671
|View full text |Cite
|
Sign up to set email alerts
|

Pop Music Transformer

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
114
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 193 publications
(114 citation statements)
references
References 14 publications
0
114
0
Order By: Relevance
“…Music generation has been widely addressed as a deep learning task ( Briot et al, 2017 ), in particular using LSTMs ( Sturm et al, 2016 ; Wu et al, 2019 ) and more recently transformers Huang et al (2018) . Music tagged with emotion has also been generated through long short-term memory networks (LSTMs) with logistic regression and used to generate music with sentiment ( Ferreira and Whitehead, 2019 ).…”
Section: Related Workmentioning
confidence: 99%
“…Music generation has been widely addressed as a deep learning task ( Briot et al, 2017 ), in particular using LSTMs ( Sturm et al, 2016 ; Wu et al, 2019 ) and more recently transformers Huang et al (2018) . Music tagged with emotion has also been generated through long short-term memory networks (LSTMs) with logistic regression and used to generate music with sentiment ( Ferreira and Whitehead, 2019 ).…”
Section: Related Workmentioning
confidence: 99%
“…A typical use of the Transformer architecture in NLP is to encode the meaning of a word given the surrounding words, sentences, and paragraphs. Beyond NLP, other example uses of the Transformer architecture are found in music generation 43 , image generation 44 , image and video restoration [45][46][47][48][49] , game playing agents 50,51 , and drug discovery 52,53 . In this work, we explore how our attention-based architecture, CrabNet, performs in predicting materials properties relative to the common modeling techniques Roost, ElemNet, and random forest (RF) for regression-type problems.…”
Section: Introductionmentioning
confidence: 99%
“…The NDT is based on the BERT encoder [5] with modifications for application to neuroscientific datasets, specifically multi-electrode spiking activity. Modifications are needed as spiking activity has markedly different statistics than both language data and other time series [8,35] previously modeled by Transformers. Further, neuroscientific datasets are generally much smaller than typical dataset sizes in other machine learning domains, necessitating careful training decisions [9] 1 .…”
Section: Introductionmentioning
confidence: 99%