2016
DOI: 10.48550/arxiv.1611.04798
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Toward Multilingual Neural Machine Translation with Universal Encoder and Decoder

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
96
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 152 publications
(97 citation statements)
references
References 8 publications
1
96
0
Order By: Relevance
“…Multilingual modeling has been an important topic in applying sequence-to-squence models to language applications ranging from machine translation [1,2] to automatic speech recognition (ASR) [3]. It is possible to employ one single neural model for multiple datasets with different languages with the goal of capturing the shared features between the languages.…”
Section: Introductionmentioning
confidence: 99%
“…Multilingual modeling has been an important topic in applying sequence-to-squence models to language applications ranging from machine translation [1,2] to automatic speech recognition (ASR) [3]. It is possible to employ one single neural model for multiple datasets with different languages with the goal of capturing the shared features between the languages.…”
Section: Introductionmentioning
confidence: 99%
“…Multilingual NMT Multilingual NMT aims to train one model to serve all language pairs (Ha et al, 2016;Firat et al, 2016;Johnson et al, 2017). Several subsequent works explored various parameter sharing strategies to mitigate the representation bottleneck (Blackwood et al, 2018;Platanios et al, 2018;Sen et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…But, due to the high complexity of the previously mentioned method, Johnson et al and Ha et al attempted to build a multilingual NMT without modifying the network architecture. Ha, Niehues, and Waibel (2016) applied a languagespecific coding to words of both source and target languages for better representation. In practice, language-specific coding for words and sub-words significantly increased the length of sentences, which causes trouble for sentence representation and attention mechanism.…”
Section: Multilingual Nmtmentioning
confidence: 99%
“…Because of the fact that NMT system is highly depended on extensive high-quality parallel data, which can only be acquired for few language pairs, it is still challenging for low-resource and zero-shot NMT (Koehn and Knowles 2017). Existing approaches for zero-shot NMT include multilingual NMT Ha, Niehues, and Waibel 2016;Johnson et al 2017), interactive multimodal framework (Kiros, Salakhutdinov, and Zemel 2014;Nakayama and Nishida 2017), pivot-based NMT (Wu and Wang 2007;Cheng et al 2016;Leng et al 2019) and teacher-student architecture (Chen et al 2017).…”
Section: Introductionmentioning
confidence: 99%