Proceedings of the Third Conference on Machine Translation: Shared Task Papers 2018
DOI: 10.18653/v1/w18-6452
|View full text |Cite
|
Sign up to set email alerts
|

Findings of the WMT 2018 Shared Task on Automatic Post-Editing

Abstract: We present the results from the fourth round of the WMT shared task on MT Automatic Post-Editing. The task consists in automatically correcting the output of a "black-box" machine translation system by learning from human corrections. Keeping the same general evaluation setting of the three previous rounds, this year we focused on one language pair (English-German) and on domainspecific data (Information Technology), with MT outputs produced by two different paradigms: phrase-based (PBSMT) and neural (NMT). Fi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
38
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 27 publications
(39 citation statements)
references
References 26 publications
1
38
0
Order By: Relevance
“…The repetition rate measures the repetitiveness inside a text by looking at the rate of non-singleton n-gram types (n=1...4) and combining them using the geometric mean. Larger values indicate a higher text repetitiveness and, as discussed in (Bojar et al, 2016;Bojar et al, 2017;Chatterjee et al, 2018a), suggest a higher chance of learning from the training set correction patterns that are applicable also to the test set. In the previous rounds of the task, we considered the large differences in repetitiveness across the datasets as a possible explanation for the variable gains over the baseline obtained by participants.…”
Section: Complexity Indicators: Repetition Ratementioning
confidence: 96%
See 1 more Smart Citation
“…The repetition rate measures the repetitiveness inside a text by looking at the rate of non-singleton n-gram types (n=1...4) and combining them using the geometric mean. Larger values indicate a higher text repetitiveness and, as discussed in (Bojar et al, 2016;Bojar et al, 2017;Chatterjee et al, 2018a), suggest a higher chance of learning from the training set correction patterns that are applicable also to the test set. In the previous rounds of the task, we considered the large differences in repetitiveness across the datasets as a possible explanation for the variable gains over the baseline obtained by participants.…”
Section: Complexity Indicators: Repetition Ratementioning
confidence: 96%
“…As discussed in (Bojar et al, 2017;Chatterjee et al, 2018a), numeric evidence of a higher quality of the original translations can indicate a smaller room for improvement for APE systems (having, at the same time, less to learn during training and less to correct at test stage). On one side, indeed, training on good (or near-perfect) automatic translations can drastically reduce the number of learned correction patterns.…”
Section: Complexity Indicators: Mt Qualitymentioning
confidence: 99%
“…However, developing a human-edit corpus is time-consuming and costly, so it cannot be performed for all language pairs. There is not a clear performance improvement when NMT results are post-edited using neural networks, unlike when SMT results are post-edited using neural networks [17].…”
Section: Related Workmentioning
confidence: 94%
“…Exploiting synthetic data has shown to lead to improvements in APE (Bojar et al 2017;Chatterjee et al 2018) and in NMT systems (Sennrich et al 2016b;Poncelas et al 2018). A detailed summary is presented in do Carmo et al (2020).…”
Section: Data Demands and Data Scarcitymentioning
confidence: 99%