2019
DOI: 10.1007/978-3-030-32233-5_29
|View full text |Cite
|
Sign up to set email alerts
|

Word Position Aware Translation Memory for Neural Machine Translation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2

Relationship

3
4

Authors

Journals

citations
Cited by 8 publications
(11 citation statements)
references
References 7 publications
0
11
0
Order By: Relevance
“…Given the input sentence x, Zhang et al (2018) try to assign target words in ŷ with higher rewards, when they appear in y r and the aligned source words are in both x r and x. He et al (2019) follow a similar framework and consider the position information of those target words when rewarding. Those works reward the target words in an explicit way, however, the one-sentence-one-model approach (Li et al, 2016c;Turchi et al, 2017) propose to reward target word implicitly.…”
Section: Inference Phasementioning
confidence: 99%
“…Given the input sentence x, Zhang et al (2018) try to assign target words in ŷ with higher rewards, when they appear in y r and the aligned source words are in both x r and x. He et al (2019) follow a similar framework and consider the position information of those target words when rewarding. Those works reward the target words in an explicit way, however, the one-sentence-one-model approach (Li et al, 2016c;Turchi et al, 2017) propose to reward target word implicitly.…”
Section: Inference Phasementioning
confidence: 99%
“…The most obvious drawback of fine-tuning is that the delay is too long for testing sentences. To avoid the online tuning process, Zhang et al (2018) and He et al (2019) dynamically integrate translation pieces, based on n-grams extracted from the matched segments in the TM target, into the beam search stage. The second type of approach is efficient but heavily depends on the global hyper-parameter λ, which is sensitive to the development set, leading to inferior performance.…”
Section: Related Workmentioning
confidence: 99%
“…Many notable approaches have been proposed to augment an NMT model by using a TM. For example, Zhang et al (2018) and He et al (2019) extract scored n-grams from a TM and then reward each partial translation once it matches an extracted n-gram during beam search. Gu et al (2018) and Xia et al (2019) use an auxiliary network to encode a TM and then integrate it into the NMT architecture.…”
Section: Introductionmentioning
confidence: 99%
“…Suppose we are given a translation memory (TM) for a source sentence which is a list of bilingual sentence pairs. Generally, there are two ways to improve translation models with translation memory: training model parameters with augmented data (i.e., memory) [41,42,43] and summarizing knowledge from translation memory to augment MT decoder [44,45,46]. For the latter idea, a typical solution to represent a TM is to encode each word in both the source and target sides by a neural memory [44].…”
Section: Graph Based Translation Memorymentioning
confidence: 99%
“…For each sentence, we retrieve 100 translation pairs from the training set by using Apache Lucene [86]. We score the source side of each retrieved pair against the source sentence with fuzzy matching score and select top N = 5 translation sentence pairs as a translation memory for the sentence to be translated, following [44,45,46]. Sentences from the target side in the translation memory are used to form a graph, with each word represented as a node and the connection between adjacent words in a sentence represented as an undirected edge.…”
Section: Translation Memorymentioning
confidence: 99%