Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1300
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Multilingual Word Embedding with Limited Resources using Neural Language Models

Abstract: We propose an unsupervised method to obtain cross-lingual embeddings without any parallel data or pre-trained word embeddings. The proposed model, which we call multilingual neural language models, takes sentences of multiple languages as an input. The proposed model contains bidirectional LSTMs that perform as forward and backward language models, and these networks are shared among all the languages. The other parameters, i.e. word embeddings and linear transformation between hidden states and outputs, are s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
32
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 33 publications
(32 citation statements)
references
References 25 publications
0
32
0
Order By: Relevance
“…We bold the highest transferring score for each pair. Multilinguality in NLP tasks This work is also related to the continual trend of multilingual language learning, including aligning multilingual word embeddings (Mikolov et al, 2013;Chen and Cardie, 2018;Lample et al, 2018b) into universal space, and learning crosslingual models (Wada and Iwata, 2018;Lample and Conneau, 2019; to exploit shared representations across languages.…”
Section: Related Workmentioning
confidence: 99%
“…We bold the highest transferring score for each pair. Multilinguality in NLP tasks This work is also related to the continual trend of multilingual language learning, including aligning multilingual word embeddings (Mikolov et al, 2013;Chen and Cardie, 2018;Lample et al, 2018b) into universal space, and learning crosslingual models (Wada and Iwata, 2018;Lample and Conneau, 2019; to exploit shared representations across languages.…”
Section: Related Workmentioning
confidence: 99%
“…Subsets of the MUSE dictionaries have been used for model comparison in the evaluation of numerous cross-lingual embedding systems developed since (cf. Jawanpuria et al, 2018;Hoshen and Wolf, 2018a,b;Wada and Iwata, 2018;. Even though the field has been very active, progress has been incremental for most language pairs.…”
Section: Introductionmentioning
confidence: 99%
“…Attempts to do so include the efforts by , who leverage an inverted index based on the Wikipedia multilingual links to generate multilingual word representations. Wada et al (2019) instead use a sentence-level neural language model for directly learning multilingual word embeddings and as a result bypassing the need for mapping functions. In the paradigm of aligning pre-trained word embeddings where we focus, Heyman et al (2019) propose a technique that iteratively builds a multilingual space starting from a monolingual space and incrementally incorporating languages to it.…”
Section: Related Workmentioning
confidence: 99%