Proceedings of the 2nd Workshop on Neural Machine Translation and Generation 2018
DOI: 10.18653/v1/w18-2702
|View full text |Cite
|
Sign up to set email alerts
|

A Shared Attention Mechanism for Interpretation of Neural Automatic Post-Editing Systems

Abstract: Automatic post-editing (APE) systems aim to correct the systematic errors made by machine translators. In this paper, we propose a neural APE system that encodes the source (src) and machine translated (mt) sentences with two separate encoders, but leverages a shared attention mechanism to better understand how the two inputs contribute to the generation of the post-edited (pe) sentences. Our empirical observations have showed that when the mt is incorrect, the attention shifts weight toward tokens in the src … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 15 publications
0
5
0
Order By: Relevance
“…At each decoder time step, an attention vector is calculated as a distribution over all encoded tokens, indicating which token the decoder should "pay attention to". To make the attention mechanism work with multiple encoders, we concatenate the hidden states of the n encoders [h 1 , ..., h n ] and apply an attention layer on the result [70].…”
Section: Core Architecturementioning
confidence: 99%
“…At each decoder time step, an attention vector is calculated as a distribution over all encoded tokens, indicating which token the decoder should "pay attention to". To make the attention mechanism work with multiple encoders, we concatenate the hidden states of the n encoders [h 1 , ..., h n ] and apply an attention layer on the result [70].…”
Section: Core Architecturementioning
confidence: 99%
“…Most of the papers looked at the application of machine translation, but there is one paper on abstractive summarization (Fan et al, 2018) and one paper on automatic post-editing of translations (Unanue et al, 2018).…”
Section: Summary Of Research Contributionsmentioning
confidence: 99%
“…Model analysis: There were also many methods that analyzed modeling and design decisions, including investigations of individual neuron contributions (Bau et al, 2018), parameter sharing (Jean et al, 2018), controlling output characteristics (Fan et al, 2018), and shared attention (Unanue et al, 2018) 3 Shared Task…”
Section: Summary Of Research Contributionsmentioning
confidence: 99%
“…shown remarkable superiority over their statistical counterparts (Varis and Bojar, 2017;Chatterjee et al, 2017;Junczys-Dowmunt and Grundkiewicz, 2017;Unanue et al, 2018). Most of them cast APE as a multi-source sequence-to-sequence learning problem (Zoph and Knight, 2016): given a source sentence (src) and a machine translation (mt), APE outputs a post-edited translation (pe).…”
Section: Introductionmentioning
confidence: 99%
“…We believe that existing approaches to APE (Varis and Bojar, 2017;Chatterjee et al, 2017;Junczys-Dowmunt and Grundkiewicz, 2017;Unanue et al, 2018) suffer from two major src:I ate a cake yesterday pe :Ich hatte gestern einen Kuchen gegessen mt :Ich esse einen Hamburger 1 0 1 0 generate copy interact Figure 1: Learning to copy for APE. Our work is based on two key ideas.…”
Section: Introductionmentioning
confidence: 99%