Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence 2018
DOI: 10.24963/ijcai.2018/357
|View full text |Cite
|
Sign up to set email alerts
|

Neural Machine Translation with Key-Value Memory-Augmented Attention

Abstract: Although attention-based Neural Machine Translation (NMT) has achieved remarkable progress in recent years, it still suffers from issues of repeating and dropping translations. To alleviate these issues, we propose a novel key-value memory-augmented attention model for NMT, called KVMEMATT. Specifically, we maintain a timely updated keymemory to keep track of attention history and a fixed value-memory to store the representation of source sentence throughout the whole translation process. Via nontrivial transf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
5

Relationship

1
9

Authors

Journals

citations
Cited by 13 publications
(5 citation statements)
references
References 0 publications
0
5
0
Order By: Relevance
“…Dialog History Memory Update: Dialog history memory update is inspired by Fandong Meng et al [19][20] , where we use both adding and forgetting to update the dialog history memory. The forget operation determines how much of the current message should be removed from the dialog history memory, and similarly, the add operation determines how much of the current message should be written to the dialog history memory as an addition.…”
Section: Dialog History Manager Modulementioning
confidence: 99%
“…Dialog History Memory Update: Dialog history memory update is inspired by Fandong Meng et al [19][20] , where we use both adding and forgetting to update the dialog history memory. The forget operation determines how much of the current message should be removed from the dialog history memory, and similarly, the add operation determines how much of the current message should be written to the dialog history memory as an addition.…”
Section: Dialog History Manager Modulementioning
confidence: 99%
“…t,j v j , where v j is the j-th memory slot in V. Dialog State Memory Updating Inspired by the read-write operations (Meng et al, 2016;Meng et al, 2018), we define two types of operations for updating the dialog state memory: FORGET and ADD. FORGET is analogous to the forget gate in GRU, which determines the information to be removed from memory slots.…”
Section: Dialog History Memory Readingmentioning
confidence: 99%
“…Tu et al (2016), Mi et al (2016) and Li et al (2018b) employ coverage vector or coverage ratio to indicate the lexical-level coverage of source words. Meng et al (2018) influence the attentive vectors by translated/untranslated information. Our work mainly follows the path of Zheng et al (2018), which introduce two extra recurrent layers in the decoder to maintain the representations of the past and future translation contents.…”
Section: Related Workmentioning
confidence: 99%