Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1671
|View full text |Cite
|
Sign up to set email alerts
|

Neural News Recommendation with Multi-Head Self-Attention

Abstract: News recommendation can help users find interested news and alleviate information overload. Precisely modeling news and users is critical for news recommendation, and capturing the contexts of words and news is important to learn news and user representations. In this paper, we propose a neural news recommendation approach with multi-head selfattention (NRMS). The core of our approach is a news encoder and a user encoder. In the news encoder, we use multi-head self-attentions to learn news representations from… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
325
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 258 publications
(325 citation statements)
references
References 19 publications
0
325
0
Order By: Relevance
“…All models mentioned above represent users and items following the same principle, and the representations are derived from the same data source. Contrary to this, NPA [43] and NeuHash-CF [12] represent users and items in different ways. While they model the items using content-based information, the users are represented by one-hot coded userID.…”
Section: Review-based User/item Modelingmentioning
confidence: 99%
“…All models mentioned above represent users and items following the same principle, and the representations are derived from the same data source. Contrary to this, NPA [43] and NeuHash-CF [12] represent users and items in different ways. While they model the items using content-based information, the users are represented by one-hot coded userID.…”
Section: Review-based User/item Modelingmentioning
confidence: 99%
“…Similarly, in the other studies a whole document or article is embedded and then its similarity with others is found [4], [53]- [57]. Moreover, It is used as base line model of feature extraction before feeding data to neural network [13], [58]- [61]. Furthermore a knowledge graph of entities are also converted in to embeddings to further use it in deep-learning settings.…”
Section: F Embeddingmentioning
confidence: 99%
“…An LDA is applied to text to get several topics, that are made input to the CNN model, which outputs the corresponding latent factor model L1. In another study [58] a convolution neural network-based news recommender system is proposed which mainly constitutes three modules, i.e., news encoder, user encoder, and click predictor. The purpose of the news encoder is to learn news representation.…”
Section: ) Cnnmentioning
confidence: 99%
“…Zhou et al [40] utilized a self-attention based sequential framework to project user representation into multiple latent spaces and models user behavior for personalized recommendation. Wu et al [41] facilitate news recommendation by modeling the contextual interaction between words and news with multi-head self-attention. Cong et al [42] distinguish the importance of reviews at both word level and sentence level using a hierarchical attention-based network for e-commerce recommendation.…”
Section: Related Workmentioning
confidence: 99%