2017
DOI: 10.48550/arxiv.1708.00107
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learned in Translation: Contextualized Word Vectors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

3
33
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 33 publications
(36 citation statements)
references
References 0 publications
3
33
0
Order By: Relevance
“…Our results are consistent with the findings in the literature -45% " 55% for in the 5-category classification, and 85% " 92% in the 2-category classification(McCann et al, 2017;Socher et al, 2013).…”
supporting
confidence: 93%
“…Our results are consistent with the findings in the literature -45% " 55% for in the 5-category classification, and 85% " 92% in the 2-category classification(McCann et al, 2017;Socher et al, 2013).…”
supporting
confidence: 93%
“…For general vision, Ima-geNet [9] pre-training can greatly assist downstream tasks, such as object detection [1,18,46] and semantic segmentation [35]. Also in natural language processing, representations pre-trained on web-crawled corpus via Mask Language Model [10] achieves leading performance on machine translation [38] and natural language inference [8].…”
Section: Related Workmentioning
confidence: 99%
“…Various early work focuses on pre-training word embeddings for downstream tasks, such as Word2Vec [24] and Glove [25]. To handle the polysemy problem of word embeddings, modern PLMs built on shallow neural networks are proposed, like CoVE [26] and ELMo [27], which can provide contextual word representations. With the introduction of Transformer [28] and the advance of distributed computing systems, PLMs built on deep neural networks have gradually appeared, such as GPT [1], BERT [3] and XLNet [2].…”
Section: Related Workmentioning
confidence: 99%