Proceedings of the Fourth Italian Conference on Computational Linguistics CLiC-it 2017 2017
DOI: 10.4000/books.aaccademia.2419
|View full text |Cite
|
Sign up to set email alerts
|

Multilingual Neural Machine Translation for Low Resource Languages

Abstract: In recent years, Neural Machine Translation (NMT) has been shown to be more effective than phrase-based statistical methods, thus quickly becoming the state of the art in machine translation (MT). However, NMT systems are limited in translating low-resourced languages, due to the significant amount of parallel data that is required to learn useful mappings between languages. In this work, we show how the so-called multilingual NMT can help to tackle the challenges associated with low-resourced language transla… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 13 publications
(6 citation statements)
references
References 13 publications
0
6
0
Order By: Relevance
“…These models use the vocabulary (word segmentation rules) of the init model, avoiding the proposed dynamic vocabulary. This fine-tuning approach is prevalent in continued model trainings, for adapting NMT models [37,38] or improving zero-shot and lowresource translation tasks [39,40,41]. For the alternative baseline where we fine-tune init with its static-vocabulary, we observed that results were mostly analogous to Bi-NMT models.…”
Section: Baseline Modelsmentioning
confidence: 84%
“…These models use the vocabulary (word segmentation rules) of the init model, avoiding the proposed dynamic vocabulary. This fine-tuning approach is prevalent in continued model trainings, for adapting NMT models [37,38] or improving zero-shot and lowresource translation tasks [39,40,41]. For the alternative baseline where we fine-tune init with its static-vocabulary, we observed that results were mostly analogous to Bi-NMT models.…”
Section: Baseline Modelsmentioning
confidence: 84%
“…The motivation behind the proposed bottleneck is the success of the Transformer neural network [45] for speech processing. A Transformer neural network reduces the dimensionality and maintains the important features of speech signals.…”
Section: A Multi-head Self Attention Transformer With Time-frequency ...mentioning
confidence: 99%
“…2012), or to aid geolocation prediction systems (Han, Cook, and Baldwin 2012). MT can address diatopic variation when translating between pairs of closely related languages (Nakov and Tiedemann 2012; Lakew, Cettolo, and Federico 2018), language varieties (Costa-jussà et al . 2018), and dialects (Zbib et al .…”
Section: Applicationsmentioning
confidence: 99%
“…2017) of neural MT, which yielded improvements over RNN-based seq2seq, for example, for Romanian–Italian and Dutch–German in bilingual, multilingual and zero-shot setups (Lakew et al . 2018).…”
Section: Applicationsmentioning
confidence: 99%