2024
DOI: 10.1609/aaai.v38i8.28747
|View full text |Cite
|
Sign up to set email alerts
|

An Attentive Inductive Bias for Sequential Recommendation beyond the Self-Attention

Yehjin Shin,
Jeongwhan Choi,
Hyowon Wi
et al.

Abstract: Sequential recommendation (SR) models based on Transformers have achieved remarkable successes. The self-attention mechanism of Transformers for computer vision and natural language processing suffers from the oversmoothing problem, i.e., hidden representations becoming similar to tokens. In the SR domain, we, for the first time, show that the same problem occurs. We present pioneering investigations that reveal the low-pass filtering nature of self-attention in the SR, which causes oversmoothing. To this end,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
references
References 37 publications
0
0
0
Order By: Relevance