Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.150
|View full text |Cite
|
Sign up to set email alerts
|

Language-aware Interlingua for Multilingual Neural Machine Translation

Abstract: Multilingual neural machine translation (NMT) has led to impressive accuracy improvements in low-resource scenarios by sharing common linguistic information across languages. However, the traditional multilingual model fails to capture the diversity and specificity of different languages, resulting in inferior performance compared with individual models that are sufficiently trained. In this paper, we incorporate a language-aware interlingua into the Encoder-Decoder architecture. The interlingual network enabl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
22
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 27 publications
(22 citation statements)
references
References 12 publications
0
22
0
Order By: Relevance
“…Existing works on translation from multiple source languages into a single low resource language usually have at most 30 source languages (Gu et al, 2018;Zhou et al, 2018a;Zhu et al, 2020). They are limited within the same or close-by language families, or those with available data, or those chosen based on the researchers' intuitive discretion.…”
Section: Ranking Source Languagesmentioning
confidence: 99%
See 1 more Smart Citation
“…Existing works on translation from multiple source languages into a single low resource language usually have at most 30 source languages (Gu et al, 2018;Zhou et al, 2018a;Zhu et al, 2020). They are limited within the same or close-by language families, or those with available data, or those chosen based on the researchers' intuitive discretion.…”
Section: Ranking Source Languagesmentioning
confidence: 99%
“…We focus on five challenges that are not addressed previously. Most multilingual transformer works that translate into low resource language limit their training data to available data in the same or close-by language families or the researchers' intuitive discretion; and are mostly limited to less than 30 languages (Gu et al, 2018;Zhou et al, 2018a;Zhu et al, 2020). Instead, we examine ways to pick useful source languages from 124 source languages in a principled fashion.…”
Section: Introductionmentioning
confidence: 99%
“…And the usage of the language token style is widely accepted. Also, many subsequent works continuously explore new approaches in MNMT, such as parameter sharing (Blackwood et al, 2018;Wang et al, 2019b;Tan et al, 2019a), parameter generation (Platanios et al, 2018), knowledge distillation (Tan et al, 2019b), learning better representation (Wang et al, 2019a), massive training (Aharoni et al, 2019;Arivazhagan et al, 2019), interlingua (Zhu et al, 2020a), and adpater (Zhu et al, 2021). These works mainly utilize parallel data.…”
Section: Multilingual Neural Machine Translationmentioning
confidence: 99%
“…Improving the consistency of semantic representations and alleviating the off-target issue (Zhang et al, 2020) are effective ways to improve the zeroshot translation quality (Al-Shedivat and Parikh, 2019;Arivazhagan et al, 2019;Zhu et al, 2020). The semantic representations of different languages should be close to each other to get better translation quality (Ding et al, 2017).…”
Section: Background and Notationsmentioning
confidence: 99%