2023
DOI: 10.4208/jml.221206
|View full text |Cite
|
Sign up to set email alerts
|

Why Self-Attention Is Natural for Sequence-to-Sequence Problems? A Perspective from Symmetries

Chao Ma Chao Ma,
Lexing Ying Lexing Ying

Abstract: In this paper, we show that structures similar to self-attention are natural for learning many sequence-to-sequence problems from the perspective of symmetry. Inspired by language processing applications, we study the orthogonal equivariance of seq2seq functions with knowledge, which are functions taking two inputs -an input sequence and a knowledge -and outputting another sequence. The knowledge consists of a set of vectors in the same embedding space as the input sequence, containing the information of the l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 39 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?