Proceedings of the Third Conference on Machine Translation: Research Papers 2018
DOI: 10.18653/v1/w18-6324
|View full text |Cite
|
Sign up to set email alerts
|

Massively Parallel Cross-Lingual Learning in Low-Resource Target Language Translation

Abstract: We work on translation from rich-resource languages to low-resource languages. The main challenges we identify are the lack of lowresource language data, effective methods for cross-lingual transfer, and the variable-binding problem that is common in neural systems. We build a translation system that addresses these challenges using eight European language families as our test ground. Firstly, we add the source and the target family labels and study intra-family and inter-family influences for effective cross-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
19
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 11 publications
(19 citation statements)
references
References 42 publications
0
19
0
Order By: Relevance
“…In their case, the source-language token is used for language specific tokenisation. Similarly, Zhou et al (2018) found that adding tokens that encode the source and target language family, e. g. source-family:Germanic and target-family:Slavic for English-Czech translation, may improve the accuracy of the NMT outputs for low-resource languages.…”
Section: Related Workmentioning
confidence: 93%
“…In their case, the source-language token is used for language specific tokenisation. Similarly, Zhou et al (2018) found that adding tokens that encode the source and target language family, e. g. source-family:Germanic and target-family:Slavic for English-Czech translation, may improve the accuracy of the NMT outputs for low-resource languages.…”
Section: Related Workmentioning
confidence: 93%
“…The two multilingual baselines are strong baselines while the fWMT baseline is a weak baseline that helps us to evaluate all experiments' performance discounting the effect of increasing data. All baseline results are taken from a research work which uses the grid of (1,6,11,16,22) for the number of languages or equivalent number of unique sentences and we follow the same in Figure 3 (Zhou et al, 2018). All experiments for each grid point carry the same number of unique sentences.…”
Section: Baselinesmentioning
confidence: 99%
“…Moreover, we have two multilingual baselines 2 built on multilingual attentional NMT, Family and Span (Zhou et al, 2018). Family refers to the multilingual baseline by adding one language family at a time, where on top of the French corpus f0 and the English corpus e0, we add up to 20 other European languages.…”
Section: Baselinesmentioning
confidence: 99%
See 1 more Smart Citation
“…In this setting, an NMT system is firstly trained using auxiliary parallel data from a so-called "parent" language pair and then the trained model is used to initialize a "child" model which is further trained on a low-resource language pair. Similar approaches that support cross-lingual transfer learning for Multi-lingual NMT train a model on the concatenation of all data instead of employing sequential training (Gu et al, 2018;Zhou et al, 2018;Wang et al, 2019).…”
Section: Introductionmentioning
confidence: 99%