Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.567
|View full text |Cite
|
Sign up to set email alerts
|

Neural Machine Translation with Monolingual Translation Memory

Abstract: Prior work has proved that Translation memory (TM) can boost the performance of Neural Machine Translation (NMT). In contrast to existing work that uses bilingual corpus as TM and employs source-side similarity search for memory retrieval, we propose a new framework that uses monolingual memory and performs learnable memory retrieval in a crosslingual manner. Our framework has unique advantages. First, the cross-lingual memory retriever allows abundant monolingual data to be TM. Second, the memory retriever an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
21
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
5

Relationship

1
9

Authors

Journals

citations
Cited by 54 publications
(42 citation statements)
references
References 34 publications
0
21
0
Order By: Relevance
“…Dense representations come from encoders, such as Transformer, trained with task-specific data. And these methods can achieve better recall performance than sparse representation on different tasks, such as open domain question answering (Karpukhin et al, 2020;Guu et al, 2020;Yu et al, 2021), knowledge-grounded generation (Zhang et al, 2021), and machine translation (Cai et al, 2021). One drawback of DPR is that it cannot process longer documents, usually less than 128 tokens (Karpukhin et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Dense representations come from encoders, such as Transformer, trained with task-specific data. And these methods can achieve better recall performance than sparse representation on different tasks, such as open domain question answering (Karpukhin et al, 2020;Guu et al, 2020;Yu et al, 2021), knowledge-grounded generation (Zhang et al, 2021), and machine translation (Cai et al, 2021). One drawback of DPR is that it cannot process longer documents, usually less than 128 tokens (Karpukhin et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Different from these tasks, keyphrase gen- eration is not a knowledge-intensive task but we treat the English passage-keyphrase training data as our knowledge. Similar approaches have been investigated in neural machine translation (Gu et al, 2018;Cai et al, 2021), dialogue (Weston et al, 2018), and knowledge-base QA (Das et al, 2021). In keyphrase generation, Chen et al (2019a); Ye et al (2021a); Kim et al (2021) retrieve similar documents from training data to produce more accurate keyphrases.…”
Section: Related Workmentioning
confidence: 94%
“…Since the decoder could learn to generate the monolingual sentences during the training, the method has been used in many studies (Edunov et al, 2018;Hoang et al, 2018;Caswell et al, 2019) to compare or improve the fluency of the model. Cai et al (2021) also used the monolingual corpus as its translation memory to retrieve the aligned target sentences. In our work, we only utilize the monolingual data to modify the train data for masking, and detach it during the translation process.…”
Section: Related Workmentioning
confidence: 99%