Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.175
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial Subword Regularization for Robust Neural Machine Translation

Abstract: Exposing diverse subword segmentations to neural machine translation (NMT) models often improves the robustness of machine translation as NMT models can experience various subword candidates. However, the diversification of subword segmentations mostly relies on the pre-trained subword language models from which erroneous segmentations of unseen words are less likely to be sampled. In this paper, we present adversarial subword regularization (ADVSR) to study whether gradient signals during training can be a su… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 30 publications
0
4
0
Order By: Relevance
“…The replacement operation of m-th token x m to the arbitrary token x can be written as δ(x m , x) := e(x) − e(x m ), where e(•) denotes embedding look-up. We induce a virtual adversarial token by the following criteria (Ebrahimi et al, 2017;Michel et al, 2019;Cheng et al, 2019;Wallace et al, 2019;Park et al, 2020):…”
Section: Gradient Informationmentioning
confidence: 99%
“…The replacement operation of m-th token x m to the arbitrary token x can be written as δ(x m , x) := e(x) − e(x m ), where e(•) denotes embedding look-up. We induce a virtual adversarial token by the following criteria (Ebrahimi et al, 2017;Michel et al, 2019;Cheng et al, 2019;Wallace et al, 2019;Park et al, 2020):…”
Section: Gradient Informationmentioning
confidence: 99%
“…Works involving Input Perturbation Apart from the works mentioned above, some works introduce subword uncertainty at the subword segmentation stage, including sampling multiple subword candidates (Kudo, 2018), applying subword dropout (Park et al, 2020) or producing adversarial subword segmentation (Provilkov et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…To tokenize words at the morpheme level, we utilize KoNLPy, an open-source library for Korean text that provides a number of different tokenizers with different parsing rules and methods. In the training process, we augmented two types of tokenized sentences from each sentence in Korean text with two different tokenizers, Mecab and Komoran, as illustrated in 10% on average, but also has the effect of subword regularization (Kudo, 2018;Park et al, 2020a). Accordingly, our model utilizes various sets of subtoken candidates, that yield robustness to typos or slangs.…”
Section: Datasetmentioning
confidence: 99%