ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9054554
|View full text |Cite
|
Sign up to set email alerts
|

Transformer VAE: A Hierarchical Model for Structure-Aware and Interpretable Music Representation Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
28
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 69 publications
(29 citation statements)
references
References 3 publications
1
28
0
Order By: Relevance
“…Furthermore, Choi et al [26] proposed a Transformerbased autoencoder that achieved global representation for the musical contexts of polyphonic piano performance data. Jiang et al [27] introduced a hierarchical Transformer VAE to learn context-sensitive melody representation with selfattention blocks, enabling the model to control the melodic and rhythmic contexts.…”
Section: B Transformer-based Music Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…Furthermore, Choi et al [26] proposed a Transformerbased autoencoder that achieved global representation for the musical contexts of polyphonic piano performance data. Jiang et al [27] introduced a hierarchical Transformer VAE to learn context-sensitive melody representation with selfattention blocks, enabling the model to control the melodic and rhythmic contexts.…”
Section: B Transformer-based Music Generationmentioning
confidence: 99%
“…To the best of our knowledge, we are the first to apply the VNMT approach to music generation. In particular, our approach is different from previous music generation studies using the variational Transformer, which mostly served as an autoencoder [26], [27].…”
Section: Introductionmentioning
confidence: 99%
“…Pop music transformer [49] adopts transformer-XL to leverage longer-range information along with a new data representation that expresses the rhythmic and harmonic structure of music. Transformer variational autoencoder [50] enables joint learning of local representation and global structure based on the hierarchical modeling. Harmony transformer [51] improves chord recognition to integrate chord segmentation with a non-autoregressive decoding method in the framework of musical harmony analysis.…”
Section: Global Structure-aware Language Modelmentioning
confidence: 99%
“…Pop music transformer [49] adopts transformer-XL to leverage longer-range information along with a new data representation that expresses the rhythmic and harmonic structure of music. Transformer variational autoencoder [50] enables the joint learning of local representation and global structure based on the hierarchical modeling. Harmony transformer [51] improves chord recognition to integrate chord segmentation with a non-autoregressive decoding method in the framework of musical harmony analysis.…”
Section: Global Structure-aware Language Modelmentioning
confidence: 99%