Proceedings of the 28th ACM International Conference on Information and Knowledge Management 2019
DOI: 10.1145/3357384.3358030
|View full text |Cite
|
Sign up to set email alerts
|

A Hierarchical Self-Attentive Model for Recommending User-Generated Item Lists

Abstract: User-generated item lists are a popular feature of many different platforms. Examples include lists of books on Goodreads, playlists on Spotify and YouTube, collections of images on Pinterest, and lists of answers on question-answer sites like Zhihu. Recommending item lists is critical for increasing user engagement and connecting users to new items, but many approaches are designed for the item-based recommendation, without careful consideration of the complex relationships between items and lists. Hence, in … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 31 publications
(11 citation statements)
references
References 34 publications
0
11
0
Order By: Relevance
“…We apply grid search for tuning the hyper-parameters of the models: the learning rate is tuned amongst {0.0001, 0.0005, 0.001, 0.005, 0.01}, the coefficient of 2 regularization is searched in {10 −5 , 10 −4 , ..., 1, 10 1 }, and the dropout ratio in {0.0, 0.1, ..., 0.5}. The set of possible hyper-parameter 7 https://pytorch.org/ values was determined on early validation tests using subsets of the datasets that we then discarded from our analyses.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We apply grid search for tuning the hyper-parameters of the models: the learning rate is tuned amongst {0.0001, 0.0005, 0.001, 0.005, 0.01}, the coefficient of 2 regularization is searched in {10 −5 , 10 −4 , ..., 1, 10 1 }, and the dropout ratio in {0.0, 0.1, ..., 0.5}. The set of possible hyper-parameter 7 https://pytorch.org/ values was determined on early validation tests using subsets of the datasets that we then discarded from our analyses.…”
Section: Methodsmentioning
confidence: 99%
“…Deep Attentive Multi-Task DAM [3] model designs a factorized attention network to aggregate the embeddings of items within a bundle to obtain the bundle's representation, while jointly model user-bundle interactions and user-item interactions in a multi-task manner to alleviate the scarcity of user-bundle interactions. Some other related efforts include [1,5,7,13,15].…”
Section: Related Workmentioning
confidence: 99%
“…In our experiments, as a non-linear network for sessions of any length, we use two-layer point-wise feed-forward networks for the item embeddings in a session and take the average of the outputs as the logits. Also we add a learnable positional embedding [3,6] to each item embedding, which encodes information about its position, in order to model the sequential pattern. That is, the encoder network for proxy selection in our experiments is built as follows:…”
Section: Implementation Detailsmentioning
confidence: 99%
“…In practice, due to the occurrence of some extreme samples, L may be set to be extremely large, resulting in the waste of memory space and affecting the speed of training. To tackle this problem, we choose the last L items that the user interacted with before the current session and put their embeddings into the long-term encoder, which is similar as in [37]. In detail, if the number of the historical items H is smaller than L, the last L − H rows of E (u,c) will be filled with 0.…”
Section: User's Long-term Preference Analysismentioning
confidence: 99%