Proceedings of the Third Conference on Machine Translation: Shared Task Papers 2018
DOI: 10.18653/v1/w18-6421
|View full text |Cite
|
Sign up to set email alerts
|

NTT’s Neural Machine Translation Systems for WMT 2018

Abstract: This paper describes NTT's neural machine translation systems submitted to the WMT 2018 English-German and German-English news translation tasks. Our submission has three main components: the Transformer model, corpus cleaning, and right-to-left nbest re-ranking techniques. Through our experiments, we identified two keys for improving accuracy: filtering noisy training sentences and right-to-left re-ranking. We also found that the Transformer model requires more training data than the RNN-based model, and the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
20
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 13 publications
(21 citation statements)
references
References 12 publications
1
20
0
Order By: Relevance
“…Further details on training are not available. (Morishita et al, 2018) NTT combine Transformer "big" model, corpus cleaning technique for provided and synthetic parallel corpora, and right-to-left n-best re-ranking techniques. Through their experiments, NTT found filtering of noisy training sentences and right-to-left re-ranking as the keys to better accuracy.…”
Section: Njunmtmentioning
confidence: 99%
“…Further details on training are not available. (Morishita et al, 2018) NTT combine Transformer "big" model, corpus cleaning technique for provided and synthetic parallel corpora, and right-to-left n-best re-ranking techniques. Through their experiments, NTT found filtering of noisy training sentences and right-to-left re-ranking as the keys to better accuracy.…”
Section: Njunmtmentioning
confidence: 99%
“…They already released earlier versions of the corpora and they were used on the WMT 2018 news shared translation tasks (Bojar et al, 2018). The WMT shared task participants reported that this corpora boosted translation accuracy when used with careful corpus cleaning (Junczys-Dowmunt, 2018;Morishita et al, 2018).…”
Section: Jparacrawl: Unconstrained Settingmentioning
confidence: 99%
“…Effectiveness of filtering noisy data in neural machine translation. Researchers in the field of neural machine translation (NMT) have recognized that collecting high-quality training data to be equally or even more important than exploring sophisticated model architectures (Koehn et al, 2018;Morishita et al, 2018). Techniques used in neural response generation and NMT are nearly identical; e.g., sequenceto-sequence models (Sutskever et al, 2014) and Transformers (Vaswani et al, 2017) are often used as base model architectures.…”
Section: Introductionmentioning
confidence: 99%