Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1321
|View full text |Cite
|
Sign up to set email alerts
|

Word and Document Embedding with vMF-Mixture Priors on Context Word Vectors

Abstract: Word embedding models typically learn two types of vectors: target word vectors and context word vectors. These vectors are normally learned such that they are predictive of some word co-occurrence statistic, but they are otherwise unconstrained. However, the words from a given language can be organized in various natural groupings, such as syntactic word classes (e.g. nouns, adjectives, verbs) and semantic themes (e.g. sports, politics, sentiment). Our hypothesis in this paper is that embedding models can be … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
2
2
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 28 publications
0
2
0
Order By: Relevance
“…Baselines. Our main baseline is the standard GloVe model, as in (1), which was found by Jameel and Schockaert (2019) to produce highly competitive entity embeddings, compared to a wide range of other methods. We also experimented with methods based on variational autoencoders, including the Neural Variational Document Model (Miao et al, 2016), but we were not able to obtain competitive results in this way.…”
Section: Methodsmentioning
confidence: 99%
“…Baselines. Our main baseline is the standard GloVe model, as in (1), which was found by Jameel and Schockaert (2019) to produce highly competitive entity embeddings, compared to a wide range of other methods. We also experimented with methods based on variational autoencoders, including the Neural Variational Document Model (Miao et al, 2016), but we were not able to obtain competitive results in this way.…”
Section: Methodsmentioning
confidence: 99%
“…Another NMF-based method was proposed by Xu et al (2018), in which the Euclidean distance was substituted with Wasserstein distance. Jameel and Schockaert (2019) rewrote the NMF objective as a cumulative product of normal distributions, in which each factor is multiplied by a von Mises-Fisher (vMF) distribution of context word vectors, to hopefully cluster the context words since the vMF density retains the cosine similarity.…”
Section: Related Workmentioning
confidence: 99%