Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2015
DOI: 10.3115/v1/n15-1157
|View full text |Cite
|
Sign up to set email alerts
|

Simple task-specific bilingual word embeddings

Abstract: We introduce a simple wrapper method that uses off-the-shelf word embedding algorithms to learn task-specific bilingual word embeddings. We use a small dictionary of easily-obtainable task-specific word equivalence classes to produce mixed context-target pairs that we use to train off-the-shelf embedding models. Our model has the advantage that it (a) is independent of the choice of embedding algorithm, (b) does not require parallel data, and (c) can be adapted to specific tasks by re-defining the equivalence … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

5
111
0

Year Published

2016
2016
2019
2019

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 102 publications
(116 citation statements)
references
References 10 publications
5
111
0
Order By: Relevance
“…Since SWTC is a less difficult task which requires coarse-grained representations, even limited amounts of training data may be sufficient to learn word embeddings which are useful for the specific task. This finding is in line with the recent work of Gouws and Søgaard (2015).…”
Section: Bwesg Vs Baseline Representationssupporting
confidence: 83%
“…Since SWTC is a less difficult task which requires coarse-grained representations, even limited amounts of training data may be sufficient to learn word embeddings which are useful for the specific task. This finding is in line with the recent work of Gouws and Søgaard (2015).…”
Section: Bwesg Vs Baseline Representationssupporting
confidence: 83%
“…Most methods rely on supervision encoded in parallel data, at the document level (Vulić and Moens, 2015), the sentence level (Zou et al, 2013;Chandar A P et al, 2014;Hermann and Blunsom, 2014;Kočiský et al, 2014;Luong et al, 2015;Coulmance et al, 2015;Oshikiri et al, 2016), or the word level (i.e. in the form of seed lexicon) (Gouws and Søgaard, 2015;Wick et al, 2016;Duong et al, 2016;Shi et al, 2015;Mikolov et al, 2013a;Faruqui and Dyer, 2014;Lu et al, 2015;Ammar et al, 2016;Zhang et al, 2016aZhang et al, , 2017Smith et al, 2017).…”
Section: Bilingual Lexicon Inductionmentioning
confidence: 99%
“…Trying to find such representations for a large multilingual vocabulary can thus become computationally prohibitive. Some attempts have recently been made in this direction, by leveraging multilingual external resources such as Wikipedia articles (Al-Rfou', Perozzi, & Skiena, 2013), or bilingual dictionaries (Gouws & Søgaard, 2015), or word-aligned parallel corpora (Klementiev, Titov, & Bhattarai, 2012), or sentence-aligned parallel corpora (Zou, Socher, Cer, & Manning, 2013;Hermann & Blunsom, 2014;Lauly, Boulanger, & Larochelle, 2014;Chandar, Lauly, Larochelle, Khapra, Ravindran, Raykar, & Saha, 2014), or document-aligned parallel corpora (Vulić & Moens, 2015). However, such external resources may not always be available for all language combinations and, when they are available (e.g., Wikipedia articles), they may be of uneven quality and quantity for languages other than English.…”
Section: Distributional Representationsmentioning
confidence: 99%