2020
DOI: 10.1007/978-3-030-61377-8_29
|View full text |Cite
|
Sign up to set email alerts
|

Deep Learning Models for Representing Out-of-Vocabulary Words

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 19 publications
0
7
0
Order By: Relevance
“…A number of challenges and limitations have been focused on, including biased data, overreliance on surface-level patterns, limited common sense, poor ability to reason and interpret feedback [506], [507]. Other issues include; the need for vast amounts of data and computational resources [508], limited generalizability [509], lack of interpretability [510], difficulty with rare or out-of-vocabulary words, limited understanding of syntax and grammar [511], and limited domainspecific knowledge [512].…”
Section: Challenges and Limitations Of Large Language Modelsmentioning
confidence: 99%
“…A number of challenges and limitations have been focused on, including biased data, overreliance on surface-level patterns, limited common sense, poor ability to reason and interpret feedback [506], [507]. Other issues include; the need for vast amounts of data and computational resources [508], limited generalizability [509], lack of interpretability [510], difficulty with rare or out-of-vocabulary words, limited understanding of syntax and grammar [511], and limited domainspecific knowledge [512].…”
Section: Challenges and Limitations Of Large Language Modelsmentioning
confidence: 99%
“…The difference between BPE and WordPiece comes mainly from how the subwords are assigned, where BPE chooses the most frequent byte pair and WordPiece chooses the the pair which maximises the likelihood of the training data. Other models try to learn to predict the meaning of an unknown word based on surrounding words, individual characters, or a combination of both (Lochter et al, 2020). Implementing an OOV solution which allows transfer learning of a pre-trained deep learning NLP encoder could potentiate more semantically accurate representations of technical language word embeddings, which in turn would improve the potential for TLS.…”
Section: Challenges and Solutionsmentioning
confidence: 99%
“…A common method to deal with OOV words, used in for instance BERT [114] and GPT [147,148,149], is to input subwords and byte-pair encodings rather than the words themselves to the model. Other models try to learn to predict the meaning of an unknown word based on surrounding words, individual characters, or a combination of both [150]. Implementing an OOV solution which allows transfer learning of a pre-trained deep learning NLP encoder could potentiate more semantically accurate representations of technical language word embeddings, which in turn would improve the potential for TLS.…”
Section: Technical Language Processingmentioning
confidence: 99%