Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.1
|View full text |Cite
|
Sign up to set email alerts
|

AligNART: Non-autoregressive Neural Machine Translation by Jointly Learning to Estimate Alignment and Translate

Abstract: Non-autoregressive neural machine translation (NART) models suffer from the multi-modality problem which causes translation inconsistency such as token repetition. Most recent approaches have attempted to solve this problem by implicitly modeling dependencies between outputs. In this paper, we introduce AligNART, which leverages full alignment information to explicitly reduce the modality of the target distribution. AligNART divides the machine translation task into (i) alignment estimation and (ii) translatio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 20 publications
(12 citation statements)
references
References 25 publications
0
12
0
Order By: Relevance
“…Gu et al [16] pre-define the latent variable Z as fertility and use it to determine how many target words every source word is aligned to. Song et al [62] predict the alignment by an aligner module as the latent variable Z. Position Information of Target Tokens.…”
Section: Latent Variable-based Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Gu et al [16] pre-define the latent variable Z as fertility and use it to determine how many target words every source word is aligned to. Song et al [62] predict the alignment by an aligner module as the latent variable Z. Position Information of Target Tokens.…”
Section: Latent Variable-based Methodsmentioning
confidence: 99%
“…FT-NAT [16] ENAT [23] NAT-REG [22] FlowSeq [57] AXE-NAT [70] Fully-NAT [38] OAXE-NAT [39] AligNART [62] DAD [68] RefineNAT [29] Insertion Transformer [53] Levenshtein [54] JM-NAT [55] Imputer [36] Multi-Task [78] RewriteNAT [26] CMLMC [56] Fully NAT Iterative NAT Fig. 9.…”
Section: Bleu Scorementioning
confidence: 99%
See 1 more Smart Citation
“…Non-autoregressive Decoding To address the inefficiency of autoregressive decoding for seq2seq generation, Gu et al (2018) first proposed non-autoregressive decoding for Machine Translation, which decodes the output sentence in one single iteration despite translation quality loss. Recent work mainly focused on improving the quality while maintaining competitive speedups, including applying various training objectives (Ghazvininejad et al, 2020a;Saharia et al, 2020;Du et al, 2021;, modeling dependencies between target tokens (Ghazvininejad et al, 2019;Qian et al, 2021;Song et al, 2021;Gu & Kong, 2021) and refining the translation outputs with multi-pass iterations (Ghazvininejad et al, 2020b;Kasai et al, 2020;Geng et al, 2021;Savinov et al, 2021;Huang et al, 2022). However, due to the inherent conditional independence assumption, non-autoregressive decoding's quality is generally less reliable than the autoregressive counterpart.…”
Section: Related Workmentioning
confidence: 99%
“…Here, the predicted full-sentence length can be considered as a latent variable during translating, aiming to help model the complex sequential dependency between incomplete source words, where introducing latent variable has been proven to provide effective help for modeling sequential dependency (Lee et al, 2018;Su et al, 2018;Shu et al, 2020;Song et al, 2021). Owing to the full-sentence length as the latent variable, the model has a stronger ability to model the sequential dependency, thereby reducing position bias.…”
Section: A Theoretical Analysis Of Position Bias In Simtmentioning
confidence: 99%