Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) 2014
DOI: 10.3115/v1/p14-2050
|View full text |Cite
|
Sign up to set email alerts
|

Dependency-Based Word Embeddings

Abstract: While continuous word embeddings are gaining popularity, current models are based solely on linear contexts. In this work, we generalize the skip-gram model with negative sampling introduced by Mikolov et al. to include arbitrary contexts. In particular, we perform experiments with dependency-based contexts, and show that they produce markedly different embeddings. The dependencybased embeddings are less topical and exhibit more functional similarity than the original skip-gram embeddings.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

20
905
0
6

Year Published

2015
2015
2018
2018

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 883 publications
(931 citation statements)
references
References 19 publications
20
905
0
6
Order By: Relevance
“…The most accurate system is WECE bow , which supports the assertion by Levy and Goldberg (2014a) that bag-of-word embeddings should offer superior performance to dependencybased embeddings on task involving semantic relations. Carrying out an error analysis, the lowest results of the WECE systems are obtained in the domains with the fewest training instances, making apparent that word embedding systems are dependent on the number of training instances.…”
Section: Domain-aware Training Instancessupporting
confidence: 75%
See 2 more Smart Citations
“…The most accurate system is WECE bow , which supports the assertion by Levy and Goldberg (2014a) that bag-of-word embeddings should offer superior performance to dependencybased embeddings on task involving semantic relations. Carrying out an error analysis, the lowest results of the WECE systems are obtained in the domains with the fewest training instances, making apparent that word embedding systems are dependent on the number of training instances.…”
Section: Domain-aware Training Instancessupporting
confidence: 75%
“…Moreover, for a fair comparison with the GraCE system, developed with dependency relations, we also tested the results obtained with a dependency-based Skip-gram model (Levy and Goldberg, 2014a). Words occurring only once in corpus are filtered out and 200-dimensional vectors are learned.…”
Section: Word Embeddings Representationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Murphy et al (2012) represented words through their co-occurrence with other words in syntactic dependency relations, and then used the Non-Negative Sparse Embedding (NNSE) method to reduce the dimension of the resulted representation. Levy and Goldberg (2014) extended the skip-gram word2vec model with negative sampling (Mikolov et al, 2013b) by basing the word co-occurrence window on the dependency parse tree of the sentence. Bollegala et al (2015) replaced bag-of-words contexts with various patterns (lexical, POS and dependency).…”
Section: Related Workmentioning
confidence: 99%
“…Traditional representation learning methods aim to capture semantic and syntactic similarities between two words [52]. A graph-based learning method is used for retrofitting word embedding by utilizing semantic lexicons [53].…”
Section: Related Workmentioning
confidence: 99%