Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1149
|View full text |Cite
|
Sign up to set email alerts
|

Deterministic Non-Autoregressive Neural Sequence Modeling by Iterative Refinement

Abstract: We propose a conditional non-autoregressive neural sequence model based on iterative refinement. The proposed model is designed based on the principles of latent variable models and denoising autoencoders, and is generally applicable to any sequence generation task. We extensively evaluate the proposed model on machine translation (En↔De and En↔Ro) and image caption generation, and observe that it significantly speeds up decoding while maintaining the generation quality comparable to the autoregressive counter… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

3
384
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 332 publications
(413 citation statements)
references
References 29 publications
3
384
0
Order By: Relevance
“…The main difficulty is converting the insert operations into in-place edits at each x i . Other parallel models (Ribeiro et al, 2018;Lee et al, 2018) have used methods like predicting insertion slots in a pre-processing step, or predicting zero or more tokens in-between any two tokens in x. We will see in Section 3.1.4 that these options do not perform well.…”
Section: The Seq2edits Functionmentioning
confidence: 99%
See 3 more Smart Citations
“…The main difficulty is converting the insert operations into in-place edits at each x i . Other parallel models (Ribeiro et al, 2018;Lee et al, 2018) have used methods like predicting insertion slots in a pre-processing step, or predicting zero or more tokens in-between any two tokens in x. We will see in Section 3.1.4 that these options do not perform well.…”
Section: The Seq2edits Functionmentioning
confidence: 99%
“…We achieve the effect of fertility using delete and append edits. Lee et al (2018) generate target sequences iteratively but require the target sequence length to be predicted at start. In contrast our in-place edit model allows target sequence length to change with appends.…”
Section: Spell Correctionmentioning
confidence: 99%
See 2 more Smart Citations
“…The lost of autoregressive dependency largely hurt the consistency of the output sentences, increase the difficulty in the learning process and thus lead to a low quality translation. Previous works mainly focus on adding different components into the NART model to improve the expressiveness of the network structure to overcome the loss of autoregressive dependency (Gu et al, 2017;Lee et al, 2018;Kaiser et al, 2018). However, the computational overhead of new components will hurt the inference speed, contradicting with the goal of the NART models: to parallelize and speed up neural machine translation models.…”
Section: Introductionmentioning
confidence: 99%