2017 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC) 2017
DOI: 10.1109/apsipa.2017.8282049
|View full text |Cite
|
Sign up to set email alerts
|

Music thumbnailing via neural attention modeling of music emotion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
12
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(13 citation statements)
references
References 7 publications
1
12
0
Order By: Relevance
“…The following observations can be made by comparing the results within the second group in Table 2: 1) LF works better than EF, 2) the positional encoding is useful for NAM, 3) with the positional encoding, NAM slightly outperforms RNAM. This validates our previous choice of LF over EF (Huang et al, 2017b), and shows that the attention mechanism can be simplified by using fully-connected layers.…”
Section: Resultssupporting
confidence: 86%
See 2 more Smart Citations
“…The following observations can be made by comparing the results within the second group in Table 2: 1) LF works better than EF, 2) the positional encoding is useful for NAM, 3) with the positional encoding, NAM slightly outperforms RNAM. This validates our previous choice of LF over EF (Huang et al, 2017b), and shows that the attention mechanism can be simplified by using fully-connected layers.…”
Section: Resultssupporting
confidence: 86%
“…We took a different approach in a previous work (Huang et al, 2017b) by utilizing the possible connections between emotions and song highlights. The idea is to firstly use a data set with emotion labels to train a neural network model for music emotion classification (Yang and Liu, 2013).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In the case of music, the author considers music enjoyment as a prediction problem based on the music that the listener has consumed before, with the hypothesis that music enjoyment consists of predicting the sounds that come next and verifying whether these expectations are met (Meyer, 1957). Huang et al (2017) use emotions to find song fragments that represent the whole song. With CNNs, the authors use emotion recognition as a way to compare if a given part of a song corresponds to the chorus section.…”
Section: Working With Human Subjectivitymentioning
confidence: 99%
“…We regard this architecture as an emotion learning model [10], which is trained over the MER31K dataset, using emotion tags from AllMusic 1 . The detail of selecting audio segments achieved by emotion learning model is shown in Fig.3.…”
Section: A Neural Attention Modelingmentioning
confidence: 99%