Proceedings of the Fourth Conference on Machine Translation (Volume 1: Research Papers) 2019
DOI: 10.18653/v1/w19-5204
|View full text |Cite
|
Sign up to set email alerts
|

APE at Scale and Its Implications on MT Evaluation Biases

Abstract: In this work, we train an Automatic Post-Editing (APE) model and use it to reveal biases in standard Machine Translation (MT) evaluation procedures. The goal of our APE model is to correct typical errors introduced by the translation process, and convert the "translationese" output into natural text. Our APE model is trained entirely on monolingual data that has been round-trip translated through English, to mimic errors that are similar to the ones introduced by NMT. We apply our model to the output of existi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
58
0
2

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 48 publications
(61 citation statements)
references
References 32 publications
1
58
0
2
Order By: Relevance
“…Our model can be regarded as an automatic postediting system -a system designed to fix systematic MT errors that is decoupled from the main MT system. Automatic post-editing has a long history, including rule-based (Knight and Chander, 1994), statistical (Simard et al, 2007) and neural approaches (Junczys-Dowmunt and Grundkiewicz, 2016;Pal et al, 2016;Freitag et al, 2019). In terms of architectures, modern approaches use neural sequence-to-sequence models, either multi-source architectures that consider both the original source and the baseline translation (Junczys-Dowmunt and Grundkiewicz, 2016;Pal et al, 2016), or monolingual repair systems, as in Freitag et al (2019), which is concurrent work to ours.…”
Section: Automatic Post-editingmentioning
confidence: 98%
See 1 more Smart Citation
“…Our model can be regarded as an automatic postediting system -a system designed to fix systematic MT errors that is decoupled from the main MT system. Automatic post-editing has a long history, including rule-based (Knight and Chander, 1994), statistical (Simard et al, 2007) and neural approaches (Junczys-Dowmunt and Grundkiewicz, 2016;Pal et al, 2016;Freitag et al, 2019). In terms of architectures, modern approaches use neural sequence-to-sequence models, either multi-source architectures that consider both the original source and the baseline translation (Junczys-Dowmunt and Grundkiewicz, 2016;Pal et al, 2016), or monolingual repair systems, as in Freitag et al (2019), which is concurrent work to ours.…”
Section: Automatic Post-editingmentioning
confidence: 98%
“…For training, the DocRepair model only requires monolingual document-level data. While we create synthetic training data via round-trip translation similarly to earlier work (Junczys-Dowmunt and Grundkiewicz, 2016;Freitag et al, 2019), note that we purposefully use sentence-level MT systems for this to create the types of consistency errors that we aim to fix with the context-aware DocRepair model. Not all types of consistency errors that we want to fix emerge from a round-trip translation, so access to parallel document-level data can be useful (Section 6.2).…”
Section: Automatic Post-editingmentioning
confidence: 99%
“…Iterative Refinement has been studied in machine translation (Lee et al, 2018;Freitag et al, 2019;Mansimov et al, 2019;Kasai et al, 2020) to gradually improve translation quality. Refinement is also used with masked language models to improve fluency of non-autoregressive generation outputs (Ghazvininejad et al, 2019;Lawrence et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…They combined artificial training data with real data provided by WMT APE tasks to train their model. Using a similar approach of generating artificial APE data, Freitag et al (2019) trained a monolingual re-writing APE model trained on the generated artificial training data alone. Contrary to the round-trip translation approach, largescale artificial APE data was generated by simply translating source sentences using NMT and SMT systems and using the reference translations as the "pseudo" post-edits to create eSCAPE corpus .…”
Section: Related Workmentioning
confidence: 99%