Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-short.47
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive Nearest Neighbor Machine Translation

Abstract: kNN-MT, recently proposed by Khandelwal et al. (2020a), successfully combines pretrained neural machine translation (NMT) model with token-level k-nearest-neighbor (kNN) retrieval to improve the translation accuracy. However, the traditional kNN algorithm used in kNN-MT simply retrieves a same number of nearest neighbors for each target token, which may cause prediction errors when the retrieved neighbors include noises. In this paper, we propose Adaptive kNN-MT to dynamically determine the number of k for eac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
45
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 37 publications
(66 citation statements)
references
References 13 publications
(9 reference statements)
1
45
0
Order By: Relevance
“…In the following, we will describe in detail the datastore construction and inference process of Policy-KNN. At the t th decoding step, we denote the retrieval results from Token-KNN as Zheng et al (2021), we construct the key vector s t using two kinds of features extracted from the retrieval results of Token-KNN. One feature is the distance, we denote the Euclidean distance between the context representation h t and i th retrieval result k tok i as d i .…”
Section: Policy-knnmentioning
confidence: 99%
See 4 more Smart Citations
“…In the following, we will describe in detail the datastore construction and inference process of Policy-KNN. At the t th decoding step, we denote the retrieval results from Token-KNN as Zheng et al (2021), we construct the key vector s t using two kinds of features extracted from the retrieval results of Token-KNN. One feature is the distance, we denote the Euclidean distance between the context representation h t and i th retrieval result k tok i as d i .…”
Section: Policy-knnmentioning
confidence: 99%
“…We apply the FAIRSEQ 3 (Ott et al 2019) toolkit for NMT implementation, and Faiss 4 (Johnson, Douze, and Jégou 2017) for efficient KNN retrieval. Following the previous experiences (Zheng et al 2021;Khandelwal et al 2020), we employ the WMT19 German-English news translation task winner model (Ng et al 2019) as the pre-trained model. The K for Token-KNN and Policy-KNN is 8.…”
Section: Experiments Experimental Setupmentioning
confidence: 99%
See 3 more Smart Citations