2017
DOI: 10.3758/s13421-017-0732-1
|View full text |Cite
|
Sign up to set email alerts
|

Estimating the average need of semantic knowledge from distributional semantic models

Abstract: Continuous bag of words (CBOW) and skip-gram are two recently developed models of lexical semantics (Mikolov, Chen, Corrado, & Dean, Advances in Neural Information Processing Systems, 26, 3111-3119, 2013). Each has been demonstrated to perform markedly better at capturing human judgments about semantic relatedness than competing models (e.g., latent semantic analysis; Landauer & Dumais, Psychological Review, 104(2), 1997 211; hyperspace analogue to language; Lund & Burgess, Behavior Research Methods, Instrumen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
34
0
1

Year Published

2019
2019
2022
2022

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 37 publications
(35 citation statements)
references
References 55 publications
0
34
0
1
Order By: Relevance
“…This network is trained on a corpus, word by word, and so the incremental development of distributional vectors over time is an inherent property of this algorithm. In the psychological literature, it has been demonstrated that these predicition-based models are mathematically equivalent to psychologically plausible learning models (Hollis, 2017;Mandera et al, 2017), such as the Rescorla-Wagner model of reinforcement learning (Rescorla & Wagner, 1972; see also Gluck & Bower, 1988;Sutton & Barto, 1981). It has also been shown that the word2vec model outperforms traditional count-based models in a variety of tasks, including the prediction of human behaviour (Baroni, Dinu, & Kruszewski, 2014;Mandera et al, 2017;Pereira, Gershman, Ritter, & Botvinick, 2016).…”
Section: Are Dsms Psychologically Implausible Learning Models?mentioning
confidence: 99%
See 1 more Smart Citation
“…This network is trained on a corpus, word by word, and so the incremental development of distributional vectors over time is an inherent property of this algorithm. In the psychological literature, it has been demonstrated that these predicition-based models are mathematically equivalent to psychologically plausible learning models (Hollis, 2017;Mandera et al, 2017), such as the Rescorla-Wagner model of reinforcement learning (Rescorla & Wagner, 1972; see also Gluck & Bower, 1988;Sutton & Barto, 1981). It has also been shown that the word2vec model outperforms traditional count-based models in a variety of tasks, including the prediction of human behaviour (Baroni, Dinu, & Kruszewski, 2014;Mandera et al, 2017;Pereira, Gershman, Ritter, & Botvinick, 2016).…”
Section: Are Dsms Psychologically Implausible Learning Models?mentioning
confidence: 99%
“…Interestingly, more recent work in the DSM literature has more thoroughly considered the question of acquisition. DSMs such as BEAGLE and word2vec have been shown to incorporate psychologically plausible learning mechanisms (Murdock, 1982;Rescorla & Wagner, 1972) to create their word representations, raising their algorithmic plausibility in comparison to earlier models (Hollis, 2017;Jones et al, 2015;Mandera et al, 2017, see the previous section). 3 Following from these arguments, DSMs should not be refuted from the outset as being mere engineering tools, as they are formulated as cognitive theories, rely on psychologically plausible assumptions, and are able to account for behavioural data.…”
Section: Are Dsms Psychologically Implausible Learning Models?mentioning
confidence: 99%
“…Thus, we can in principle derive language-based representation for all words occurring in a language via the cbow model. In several works, the cbow model has been identified as a psychologically plausible model for the acquisition of semantic representations (Hollis, 2017;Mandera et al, 2017). In order to obtain reliable word embeddings, we only considered words with a frequency larger than 50 in the source corpus.…”
Section: Language-based Representationsmentioning
confidence: 99%
“…This algorithm trains distributional vectors as the hidden layer of a neural network predicting the n words surrounding a target word to each side (see Hollis, 2017;Mandera, Keuleers, & Brysbaert, 2017;Mikolov et al, 2013, for technical details). This model has SEMANTIC TRANSPARENCY OF GERMAN COMPOUNDS 11 been shown to implement psychologically plausible learning mechanisms for semantic representations (Hollis, 2017;Mandera et al, 2017).…”
Section: Semantic Model and Measuresmentioning
confidence: 99%