2007
DOI: 10.3758/bf03193020
|View full text |Cite
|
Sign up to set email alerts
|

Extracting semantic representations from word co-occurrence statistics: A computational study

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

14
597
3
8

Year Published

2013
2013
2018
2018

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 582 publications
(622 citation statements)
references
References 31 publications
14
597
3
8
Order By: Relevance
“…Fyshe, Talukdar, Murphy, and Mitchell (2013) report peak performance with around 600 dimensions. Bullinaria and Levy (2007) report maximum performance around 1,000 dimensions. The authors of these two studies used PPMI instead of a raw co-occurrence frequency prior to dimension reduction and argued that PPMI may capture more semantic information in these higher dimensions than raw cooccurrence frequencies.…”
Section: Evaluation Of Semantic Space Modelsmentioning
confidence: 99%
See 3 more Smart Citations
“…Fyshe, Talukdar, Murphy, and Mitchell (2013) report peak performance with around 600 dimensions. Bullinaria and Levy (2007) report maximum performance around 1,000 dimensions. The authors of these two studies used PPMI instead of a raw co-occurrence frequency prior to dimension reduction and argued that PPMI may capture more semantic information in these higher dimensions than raw cooccurrence frequencies.…”
Section: Evaluation Of Semantic Space Modelsmentioning
confidence: 99%
“…Additionally, in the HAL model multidimensional scaling was performed on a different set of semantic vectors from three different, superordinate, semantic categories, and the results indicated that the same vectors contain Bcategorical^semantic information. These and a number of other word similarity judgment tasks have remained common benchmarks for assessing the quality of semantic space models (Bullinaria & Levy, 2007;Fyshe et al, 2013;Rohde, Gonnerman, & Plaut, 2006). A number of similarity metrics have also been used for comparing vectors in semantic spaces; the Euclidean distance and the cosine similarity are among the most popular (see Rohde et al, 2006, for a detailed discussion).…”
Section: Evaluation Of Semantic Space Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…Concordances are probably the most simple scheme to examine contextual semantic effects, but leave semantic inferences entirely to the human observer. A more complex layer is reached with collocations which can be identified automatically via statistical word co-occurrence metrics (Manning and Schütze, 1999;Wermter and Hahn, 2006), two of which are incorporated in JESEME as well: Positive pointwise mutual information (PPMI), developed by Bullinaria and Levy (2007) as an improvement over the probability ratio of normal pointwise mutual information (PMI; Church and Hanks (1990)) and Pearson's χ 2 , commonly used for testing the association between categorical variables (e.g., POS tags) and considered to be more robust than PMI when facing sparse information (Manning and Schütze, 1999).…”
Section: Distributional Semanticsmentioning
confidence: 99%