2020
DOI: 10.1609/aaai.v34i05.6418
|View full text |Cite
|
Sign up to set email alerts
|

Alignment-Enhanced Transformer for Constraining NMT with Pre-Specified Translations

Abstract: We investigate the task of constraining NMT with pre-specified translations, which has practical significance for a number of research and industrial applications. Existing works impose pre-specified translations as lexical constraints during decoding, which are based on word alignments derived from target-to-source attention weights. However, multiple recent studies have found that word alignment derived from generic attention heads in the Transformer is unreliable. We address this problem by introducing a de… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
44
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 39 publications
(44 citation statements)
references
References 16 publications
0
44
0
Order By: Relevance
“…In addition to AER, we compare the performance of NAIVE-ATT, SHIFT-ATT and SHIFT-AET on dictionary-guided machine translation (Song et al, 2020), which is an alignment-based downstream task. Given source and target constraint pairs from dictionary, the NMT model is encouraged to translate with provided constraints via word alignments (Alkhouli et al, 2018;Hasler et al, 2018;Hokamp and Liu, 2017;Song et al, 2020). More specifically, at each decoding step, the last token of the candidate translation will be revised with target constraint if it is aligned to the corresponding source constraint according to the alignment induction method.…”
Section: Downstream Task Resultsmentioning
confidence: 99%
“…In addition to AER, we compare the performance of NAIVE-ATT, SHIFT-ATT and SHIFT-AET on dictionary-guided machine translation (Song et al, 2020), which is an alignment-based downstream task. Given source and target constraint pairs from dictionary, the NMT model is encouraged to translate with provided constraints via word alignments (Alkhouli et al, 2018;Hasler et al, 2018;Hokamp and Liu, 2017;Song et al, 2020). More specifically, at each decoding step, the last token of the candidate translation will be revised with target constraint if it is aligned to the corresponding source constraint according to the alignment induction method.…”
Section: Downstream Task Resultsmentioning
confidence: 99%
“…For downstream tasks, word alignment can be used to improve dictionary-guided NMT (Song et al, 2020;Chen et al, 2020). Specifically, at each decoding step in NMT, Chen et al ( 2020) used a SHIFT-AET method to compute word alignment for the newly generated target word and then revised the newly generated target word by encouraging the pre-specified translation from the dictionary.…”
Section: Dictionary-guided Nmt Via Word Alignmentmentioning
confidence: 99%
“…Previously, some NMT with terminology constraints have been studied (Hasler et al, 2018;Alkhouli et al, 2018;Dinu et al, 2019;Chen et al, 2020;Song et al, 2020). For example, Song et al (2020) proposed a dedicated head in a multi-head Transformer architecture to learn explicit word alignment and use it to guide the constrained decoding process. When the source-aligned word matches a dictionary, the model outputs the corresponding target word.…”
Section: Related Workmentioning
confidence: 99%
“…Since the emergence of neural machine translation (NMT) models (Sutskever et al, 2014;Bahdanau et al, 2015;Vaswani et al, 2017), several studies have been conducted to explore NMT systems capable of decoding translations under terminological constraints (Hasler et al, 2018;Dinu et al, 2019;Chen et al, 2020;Song et al, 2020). However, these previous studies were conducted under the condition that a bilingual dictionary is given.…”
Section: Introductionmentioning
confidence: 99%