Proceedings of the First Conference on Machine Translation: Volume 2, Shared Task Papers 2016
DOI: 10.18653/v1/w16-2378
|View full text |Cite
|
Sign up to set email alerts
|

Log-linear Combinations of Monolingual and Bilingual Neural Machine Translation Models for Automatic Post-Editing

Abstract: This paper describes the submission of the AMU (Adam Mickiewicz University) team to the Automatic Post-Editing (APE) task of WMT 2016. We explore the application of neural translation models to the APE problem and achieve good results by treating different models as components in a log-linear model, allowing for multiple inputs (the MT-output and the source) that are decoded to the same target language (post-edited translations). A simple string-matching penalty integrated within the log-linear model is used t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
29
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 69 publications
(29 citation statements)
references
References 16 publications
0
29
0
Order By: Relevance
“…Our model can be regarded as an automatic postediting system -a system designed to fix systematic MT errors that is decoupled from the main MT system. Automatic post-editing has a long history, including rule-based (Knight and Chander, 1994), statistical (Simard et al, 2007) and neural approaches (Junczys-Dowmunt and Grundkiewicz, 2016;Pal et al, 2016;Freitag et al, 2019). In terms of architectures, modern approaches use neural sequence-to-sequence models, either multi-source architectures that consider both the original source and the baseline translation (Junczys-Dowmunt and Grundkiewicz, 2016;Pal et al, 2016), or monolingual repair systems, as in Freitag et al (2019), which is concurrent work to ours.…”
Section: Automatic Post-editingmentioning
confidence: 98%
See 1 more Smart Citation
“…Our model can be regarded as an automatic postediting system -a system designed to fix systematic MT errors that is decoupled from the main MT system. Automatic post-editing has a long history, including rule-based (Knight and Chander, 1994), statistical (Simard et al, 2007) and neural approaches (Junczys-Dowmunt and Grundkiewicz, 2016;Pal et al, 2016;Freitag et al, 2019). In terms of architectures, modern approaches use neural sequence-to-sequence models, either multi-source architectures that consider both the original source and the baseline translation (Junczys-Dowmunt and Grundkiewicz, 2016;Pal et al, 2016), or monolingual repair systems, as in Freitag et al (2019), which is concurrent work to ours.…”
Section: Automatic Post-editingmentioning
confidence: 98%
“…For training, the DocRepair model only requires monolingual document-level data. While we create synthetic training data via round-trip translation similarly to earlier work (Junczys-Dowmunt and Grundkiewicz, 2016;Freitag et al, 2019), note that we purposefully use sentence-level MT systems for this to create the types of consistency errors that we aim to fix with the context-aware DocRepair model. Not all types of consistency errors that we want to fix emerge from a round-trip translation, so access to parallel document-level data can be useful (Section 6.2).…”
Section: Automatic Post-editingmentioning
confidence: 99%
“…In QE task, various methods are proposed, such as QuEst++, 10 which is a typical baseline using many features, a method that uses Gaussian Processes, 11 a method that predicts human post-edited sentences from original and translated sentences 12,13 and a Predictor-Estimator model [14][15][16] whose Predictor extracts features from source and translated sentences and Estimator estimates word tags and edit distances based on extracted features. Recently, many methods using pre-trained language models such as BERT and ELMo have been proposed.…”
Section: Using Post-edited Translations In Training a Quality Estimatmentioning
confidence: 99%
“…This means more careful decisions have to be made by the APE system, making the least possible edits to the raw mt. To this aim, we introduce our "conservativeness" penalty developed on the post editing penalty proposed by (Junczys-Dowmunt and Grundkiewicz, 2016). It is a simple yet effective method to penalize/reward hypotheses in the beam, at inference time, that diverge far from the original input.…”
Section: Conservativeness Penaltymentioning
confidence: 99%