2016
DOI: 10.1613/jair.4986
|View full text |Cite
|
Sign up to set email alerts
|

Bilingual Distributed Word Representations from Document-Aligned Comparable Data

Abstract: We propose a new model for learning bilingual word representations from non-parallel document-aligned data. Following the recent advances in word representation learning, our model learns dense real-valued word vectors, that is, bilingual word embeddings (BWEs). Unlike prior work on inducing BWEs which heavily relied on parallel sentence-aligned corpora and/or readily available translation resources such as dictionaries, the article reveals that BWEs may be learned solely on the basis of document-aligned compa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
87
0
1

Year Published

2017
2017
2020
2020

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 83 publications
(88 citation statements)
references
References 70 publications
0
87
0
1
Order By: Relevance
“…The first variant of the architecture assumes that W S and W T are obtained in advance using any state-of-theart word embedding model, e.g., (Mikolov et al, 2013b;Vulić and Moens, 2016). They are then kept fixed when minimizing the loss from Eq.…”
Section: Word-level Encodermentioning
confidence: 99%
See 3 more Smart Citations
“…The first variant of the architecture assumes that W S and W T are obtained in advance using any state-of-theart word embedding model, e.g., (Mikolov et al, 2013b;Vulić and Moens, 2016). They are then kept fixed when minimizing the loss from Eq.…”
Section: Word-level Encodermentioning
confidence: 99%
“…To test the generality of our approach, we experiment with two well-known embedding models: (1) the model from Mikolov et al (2013b), which trains monolingual embeddings using skipgram with negative sampling (SGNS) (Mikolov et al, 2013a); and (2) the model of Vulić and Moens (2016) which learns word-level bilingual embeddings from document-aligned comparable data (BWESG). For both models, the top layers of our proposed classification network should learn to relate the word-level features stemming from these word embeddings using a set of annotated translation pairs.…”
Section: Word-level Encodermentioning
confidence: 99%
See 2 more Smart Citations
“…A possible relaxation is to use document-aligned or label-aligned comparable corpora (Søgaard et al, 2015;Vulić and Moens, 2016;Mogadala and Rettinger, 2016), but large amounts of such corpora are not always available for some language pairs.…”
Section: Introductionmentioning
confidence: 99%