Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.153
|View full text |Cite
|
Sign up to set email alerts
|

Crafting Adversarial Examples for Neural Machine Translation

Abstract: Effective adversary generation for neural machine translation (NMT) is a crucial prerequisite for building robust machine translation systems. In this work, we investigate veritable evaluations of NMT adversarial attacks, and propose a novel method to craft NMT adversarial examples. We first show the current NMT adversarial attacks may be improperly estimated by the commonly used monodirectional translation, and we propose to leverage the round-trip translation technique to build valid metrics for evaluating N… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
18
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 30 publications
(18 citation statements)
references
References 17 publications
0
18
0
Order By: Relevance
“…Even though the success rate of concatenation attack lags behind the state-of-the-art textual attack, the manipulation attack achieves performance of the same ballpark, which demonstrates the efficacy of optimization-based attack and our solvers. More importantly, it implies that the attack is not transferable between the two tasks, documenting more evidence on language attack transferability (Yuan et al, 2021;He et al, 2021). The bottom line is that they are two different tasks under different assumptions.…”
Section: Resultsmentioning
confidence: 91%
See 1 more Smart Citation
“…Even though the success rate of concatenation attack lags behind the state-of-the-art textual attack, the manipulation attack achieves performance of the same ballpark, which demonstrates the efficacy of optimization-based attack and our solvers. More importantly, it implies that the attack is not transferable between the two tasks, documenting more evidence on language attack transferability (Yuan et al, 2021;He et al, 2021). The bottom line is that they are two different tasks under different assumptions.…”
Section: Resultsmentioning
confidence: 91%
“…Our code is available at https://github.com/yonxie/ AdvFinTweet It is now known that text-based deep learning models can be vulnerable to adversarial attacks (Szegedy et al, 2014;Goodfellow et al, 2015). The perturbation can be at the sentence level (e.g., Xu et al, 2021;Iyyer et al, 2018;Ribeiro et al, 2018), the word level (e.g., Zhang et al, 2019;Alzantot et al, 2018;Zang et al, 2020;Jin et al, 2020;Lei et al, 2019;Lin et al, 2021), or both (Chen et al, 2021). We are interested in whether such adversarial attack vulnerability also exists in stock prediction models, as these models embrace more and more human-generated media data (e.g., Twitter, Reddit, Stocktwit, Yahoo News (Xu and Cohen, 2018;Sawhney et al, 2021)).…”
Section: Introductionmentioning
confidence: 99%
“…The previous approaches for constructing NMT adversarial examples can be divided into two branches: white-box and black-box. The whitebox approaches are based on the assumption that the architecture and parameters of the NMT model are accessible (Ebrahimi et al, 2018;Cheng et al, 2019;Chen et al, 2021). These methods usually achieve superior performance since they can construct and defend the adversaries tailored for the model.…”
Section: Adversarial Examples For Nmtmentioning
confidence: 99%
“…Our code is available at https://github.com/yonxie/ AdvFinTweet It is now known that text-based deep learning models can be vulnerable to adversarial attacks (Szegedy et al, 2014;Goodfellow et al, 2015). The perturbation can be at the sentence level (e.g., Xu et al, 2021;Iyyer et al, 2018;Ribeiro et al, 2018), the word level (e.g., Zhang et al, 2019;Alzantot et al, 2018;Zang et al, 2020;Jin et al, 2020;Lei et al, 2019;Zhang et al, 2021;Lin et al, 2021), or both (Chen et al, 2021). We are interested in whether such adversarial attack vulnerability also exists in stock prediction models, as these models embrace more and more human-generated media data (e.g., Twitter, Reddit, Stocktwit, Yahoo News (Xu and Cohen, 2018;Sawhney et al, 2021)).…”
Section: Introductionmentioning
confidence: 99%