Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2016
DOI: 10.18653/v1/p16-1024
|View full text |Cite
|
Sign up to set email alerts
|

On the Role of Seed Lexicons in Learning Bilingual Word Embeddings

Abstract: A shared bilingual word embedding space (SBWES) is an indispensable resource in a variety of cross-language NLP and IR tasks. A common approach to the SB-WES induction is to learn a mapping function between monolingual semantic spaces, where the mapping critically relies on a seed word lexicon used in the learning process. In this work, we analyze the importance and properties of seed lexicons for the SBWES induction across different dimensions (i.e., lexicon source, lexicon size, translation method, translati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

6
96
1

Year Published

2016
2016
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 80 publications
(103 citation statements)
references
References 33 publications
6
96
1
Order By: Relevance
“…• The method Ridge is used in a number of prior work (Mikolov et al, 2013;Dinu et al, 2014;Vulic and Korhonen, 2016). These approaches use an L2-regularized least-squares error objective as shown in Equation 2.…”
Section: Methods Under Comparisonmentioning
confidence: 99%
See 3 more Smart Citations
“…• The method Ridge is used in a number of prior work (Mikolov et al, 2013;Dinu et al, 2014;Vulic and Korhonen, 2016). These approaches use an L2-regularized least-squares error objective as shown in Equation 2.…”
Section: Methods Under Comparisonmentioning
confidence: 99%
“…Our work is most related to models that do zero-shot learning for bilingual dictionary induction, using maps between vector spaces with seed dictionaries as training data. Examples include the models of (Mikolov et al, 2013;Dinu et al, 2014;Lazaridou et al, 2015;Vulic and Korhonen, 2016). Like these approaches, we first learn word embeddings for each language, then use a seed dictionary to train a mapping function between the two vector spaces.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…More recently, Vulic et al [8] presented a systematic study of four classes of methods for learning bilingual embeddings. The authors find approach based on linear projection, similar to the one we use in our method, to be most practical and efficient.…”
Section: Related Workmentioning
confidence: 99%