Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 1 2017
DOI: 10.18653/v1/e17-1100
|View full text |Cite
|
Sign up to set email alerts
|

A Multifaceted Evaluation of Neural versus Phrase-Based Machine Translation for 9 Language Directions

Abstract: We aim to shed light on the strengths and weaknesses of the newly introduced neural machine translation paradigm. To that end, we conduct a multifaceted evaluation in which we compare outputs produced by state-of-the-art neural machine translation and phrase-based machine translation systems for 9 language directions across a number of dimensions. Specifically, we measure the similarity of the outputs, their fluency and amount of reordering, the effect of sentence length and performance across different error … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

7
54
0
3

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 79 publications
(64 citation statements)
references
References 30 publications
7
54
0
3
Order By: Relevance
“…Furthermore, NMT's reorderings are closer to the reorderings in the reference than those of PB-SMT (Toral and Sánchez-Cartagena, 2017). Thus, in this paper, we ask the question whether pre-reordering is necessary and helpful for NMT.…”
Section: Introductionmentioning
confidence: 94%
See 1 more Smart Citation
“…Furthermore, NMT's reorderings are closer to the reorderings in the reference than those of PB-SMT (Toral and Sánchez-Cartagena, 2017). Thus, in this paper, we ask the question whether pre-reordering is necessary and helpful for NMT.…”
Section: Introductionmentioning
confidence: 94%
“…The stateof-the-art NMT model employs an encoder-decoder architecture with an attention mechanism, in which the encoder summarizes the source sentence into a vector representation, and the decoder produces the target string word by word from vector representations, and the attention mechanism learns the soft alignment of a target word against source words (Bahdanau et al, 2015). NMT systems have outperformed the state-of-the-art SMT model on various language pairs in terms of translation qual-ity (Luong et al, 2015;Bentivogli et al, 2016;Junczys-Dowmunt et al, 2016;Wu et al, 2016;Toral and Sánchez-Cartagena, 2017). However, due to some deficiencies of NMT systems such as the limited vocabulary size, low adequacy for some translations, much research work has involved incorporating extra knowledge such as SMT features or linguistic features into NMT to improve translation performance (He et al, 2016;Sennrich and Haddow, 2016;Nadejde et al, 2017;Wang et al, 2017).…”
Section: Introductionmentioning
confidence: 99%
“…Bentivogli et al (2016) report that English-German NMT post-editing was reduced on average by 26% when compared with the best-performing SMT system, with fewer word order, lexical, and morphological errors, concluding that NMT has "significantly pushed ahead the state of the art", particularly for morphologically rich languages. Toral and Sánchez-Cartagena (2017) compare NMT and PBSMT for nine language pairs (English to and from Czech, German, Romanian, Russian, and English to Finnish), with engines trained for the WMT newstest data. Better automatic evaluation results are obtained for NMT output than for PBSMT output for all language pairs other than Russian-English and Romanian-English.…”
Section: The Rise Of Neural Machine Translation Modelsmentioning
confidence: 99%
“…They conducted automatic analysis on manually post-edited data in terms of morphological, lexical and ordering errors together with the fine grained analysis of ordering errors and found out that the main advantage of NMT approach is better ordering, especially for verbs. (Toral and Sánchez-Cartagena, 2017) performed a multifaceted automatic analysis based on independent human reference translations for nine language pairs from news domain. The analysis consists of output similarity, fluency measured by LM perplexity, degree of reordering as well as three broad error classes: morphological, reordering and lexical errors.…”
Section: Related Workmentioning
confidence: 99%
“…(Bentivogli et al, 2016) conducted a detailed analysis for the English-to-German translation of transcribed TED talks and found out that NMT (i) decreases post-editing effort, (ii) degrades faster than PBMT with sentence length and (iii) results in a notable improvement regarding reordering, especially for verbs. (Toral and Sánchez-Cartagena, 2017) go further in this direction by conducting a multilingual and multifaceted evaluation and found out that (i) NMT outputs are considerably different than PBMT outputs, (ii) NMT outputs are more fluent, (iii) NMT systems introduce more reorderings than PBMT systems, (iv) PBMT outperforms NMT for very long sentences and (v) NMT performs better in terms of morphological and reordering errors across all language pairs.…”
Section: Introductionmentioning
confidence: 99%