2017
DOI: 10.48550/arxiv.1702.02170
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

How to evaluate word embeddings? On importance of data efficiency and simple supervised tasks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(19 citation statements)
references
References 0 publications
0
18
0
Order By: Relevance
“…GloVe is trained for 100 epochs and parameters are set to those recommended in the paper (x-max is 100 and context window size is 10). We used open-source word embedding evaluation tool kits for assessing the quality of the models (Jastrzebski et al, 2017;Conneau and Kiela, 2018). Table 4: Interpretation of attention.…”
Section: Methodsmentioning
confidence: 99%
“…GloVe is trained for 100 epochs and parameters are set to those recommended in the paper (x-max is 100 and context window size is 10). We used open-source word embedding evaluation tool kits for assessing the quality of the models (Jastrzebski et al, 2017;Conneau and Kiela, 2018). Table 4: Interpretation of attention.…”
Section: Methodsmentioning
confidence: 99%
“…The word pairs are selected from the following benchmark datasets: RG (Rubenstein and Goodenough, 1965), MTurk (Radinsky et al, 2011), RW (Luong et al, 2013), MEN (Bruni et al, 2014), SimLex999 (Hill et al, 2015) and AP (Almuhareb and Poesio, 2005). The results for these tests are obtained from the word embedding benchmark package (Jastrzebski et al, 2017) 7 . Note that it is not our primary aim to achieve a state-of-the-art result in this test.…”
Section: Word Semantic Similarity Testmentioning
confidence: 99%
“…Many benchmarks have been proposed for the evaluation of unsupervised word representations. In general, they can be divided into intrinsic and extrinsic evaluation methods (Schnabel et al, 2015;Chiu et al, 2016;Jastrzebski et al, 2017;Alshargi et al, 2018;Bakarov, 2018). While most datasets report the semantic similarity between words, many datasets actually capture semantic relatedness (Hill et al, 2015;Avraham and Goldberg, 2016), or more complex relations such as analogy or the ability to categorize words based on the distributed representation encoded in word embeddings.…”
Section: Word Similarity and Relatednessmentioning
confidence: 99%