Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1257
|View full text |Cite
|
Sign up to set email alerts
|

Investigating Different Syntactic Context Types and Context Representations for Learning Word Embeddings

Abstract: The number of word embedding models is growing every year. Most of them are based on the co-occurrence information of words and their contexts. However, it is still an open question what is the best definition of context. We provide a systematical investigation of 4 different syntactic context types and context representations for learning word embeddings. Comprehensive experiments are conducted to evaluate their effectiveness on 6 extrinsic and intrinsic tasks. We hope that this paper, along with the publishe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
25
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 27 publications
(26 citation statements)
references
References 36 publications
1
25
0
Order By: Relevance
“…• Subjectivity/objectivity classification: Rotten Tomato snippets (Pang and Lee, 2004), using a logistic regression over summed word embeddings (Li et al, 2017a).…”
Section: Discussionmentioning
confidence: 99%
“…• Subjectivity/objectivity classification: Rotten Tomato snippets (Pang and Lee, 2004), using a logistic regression over summed word embeddings (Li et al, 2017a).…”
Section: Discussionmentioning
confidence: 99%
“…We follow the evaluation protocol for sequential labeling used by Kiros et al (2015) and Li et al (2017), and use logistic regression classifier 13 as the model for POS tagging. When predicting the tag for the i-th word w i in a sentence, the input to the classifier is the concatenation of the vectors w i−2 , w i−1 , w i , w i+1 , w i+2 for the word itself and the words in its context.…”
Section: Pos Tagging Modelmentioning
confidence: 99%
“…For example, using California as a negative sample for Oregon helps the model to learn that the pattern "X is located in Y" fits the pair (Portland, Oregon), but not the pair (Portland, California). Similar adversarial constraints were used in knowledge base completion (Toutanova et al, 2015) and word embeddings (Li et al, 2017). 4…”
Section: Objectivementioning
confidence: 99%