Proceedings of the Canadian Conference on Artificial Intelligence 2021
DOI: 10.21428/594757db.37d7654d
|View full text |Cite
|
Sign up to set email alerts
|

Position Encoding Schemes for Linear Aggregation of Word Sequences

Abstract: Deep Averaging Networks (DANs) show strong performance in several key Natural Language Processing (NLP) tasks. However, their chief drawback is not accounting for the position of tokens when encoding sequences. We study how existing position encodings might be integrated into the DAN architecture. In addition, we propose a novel position encoding built specifically for DANs, which allows greater generalization capabilities to unseen lengths of sequences. This is demonstrated on decision tasks on binary sequenc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 7 publications
0
1
0
Order By: Relevance
“…The former is provided by a symbol embedding (e.g., pretrained word vectors) of dimension m, e(w i ) ∈ R m . The latter is based on a positional encoding inspired by Maupomé et al [12]. This positional encoding is as follows: given a vector of parameters, s ∈ R m , the jth component of the encoding for position i, p(i), is given by:…”
Section: Contextualizermentioning
confidence: 99%
“…The former is provided by a symbol embedding (e.g., pretrained word vectors) of dimension m, e(w i ) ∈ R m . The latter is based on a positional encoding inspired by Maupomé et al [12]. This positional encoding is as follows: given a vector of parameters, s ∈ R m , the jth component of the encoding for position i, p(i), is given by:…”
Section: Contextualizermentioning
confidence: 99%