Proceedings of the Tenth Workshop on Statistical Machine Translation 2015
DOI: 10.18653/v1/w15-3027
|View full text |Cite
|
Sign up to set email alerts
|

Why Predicting Post-Edition is so Hard? Failure Analysis of LIMSI Submission to the APE Shared Task

Abstract: This paper describes the two systems submitted by LIMSI to the WMT'15 Shared Task on Automatic Post-Editing. The first one relies on a reformulation of the APE task as a Machine Translation task; the second implements a simple rule-based approach. Neither of these two systems manage to improve the automatic translation. We show, by carefully analyzing the failure of our systems that this counterperformance mainly results from the inconsistency in the annotations.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 3 publications
0
6
0
Order By: Relevance
“…The higher is the value of the negimpact feature, the less useful is the translation option. After pruning, the final stage of the APE pipeline tries to raise the capability of the decoder to select the correct translation rule by the introduction of new task specific features integrated in (Wisniewski et al, 2015) USAAR-SAPE Saarland University, Germany & Jadavpur University, India (Pal et al, 2015b) the model. These features measure the similarity and the reliability of the translation options and help to improve the precision of the resulting APE system.…”
Section: Participantsmentioning
confidence: 99%
See 1 more Smart Citation
“…The higher is the value of the negimpact feature, the less useful is the translation option. After pruning, the final stage of the APE pipeline tries to raise the capability of the decoder to select the correct translation rule by the introduction of new task specific features integrated in (Wisniewski et al, 2015) USAAR-SAPE Saarland University, Germany & Jadavpur University, India (Pal et al, 2015b) the model. These features measure the similarity and the reliability of the translation options and help to improve the precision of the resulting APE system.…”
Section: Participantsmentioning
confidence: 99%
“…For the first edition of the APE shared task LIMSI submitted two systems (Wisniewski et al, 2015). The first one is based on the approach of Simard et al (2007) and considers the APE task as a monolingual translation between a translation hypothesis and its post-edition.…”
Section: Limsimentioning
confidence: 99%
“…The team from FBK (Chatterjee et al 2015a) addressed data sparsity (most of the entries in the MT phrase table are unique) with a feature that measured the usefulness of each translated unit and pruned away the least useful ones. The team from LIMSI (Wisniewski et al 2015) tried to develop sieves of rules that tackled known grammatical issues. Finally, Pal et al (2015) tested different phrase lengths for the language and the translation model.…”
Section: Wmt 2015: the Stone Age Of Apementioning
confidence: 99%
“…We may go back to the pilot APE shared task in 2015 to see a detailed analysis of data issues beyond the repetitiveness of the training data. Wisniewski et al (2015) presented a system that tried to learn edits by using edit distances. Besides the overcorrection effect mentioned earlier, the authors mention "uniqueness of edits" as a major issue: even the most frequent edits (e.g.…”
Section: Data Issuesmentioning
confidence: 99%
“…Other popular approaches rely on rule-based components (Wisniewski et al, 2015;Béchara et al, 2012) which we do not discuss here.…”
Section: Post-editingmentioning
confidence: 99%