2021
DOI: 10.48550/arxiv.2107.11275
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Differentiable Language Model Adversarial Attack on Text Classifiers

Abstract: Robustness of huge Transformer-based models for natural language processing is an important issue due to their capabilities and wide adoption. One way to understand and improve robustness of these models is an exploration of an adversarial attack scenario: check if a small perturbation of an input can fool a model.Due to the discrete nature of textual data, gradient-based adversarial methods, widely used in computer vision, are not applicable per se. The standard strategy to overcome this issue is to develop t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 27 publications
0
3
0
Order By: Relevance
“…All rights reserved. 2019; Chen et al 2021b), text classification (Fursov et al 2021), and speaker recognition (Chen et al 2021a).…”
Section: Introductionmentioning
confidence: 99%
“…All rights reserved. 2019; Chen et al 2021b), text classification (Fursov et al 2021), and speaker recognition (Chen et al 2021a).…”
Section: Introductionmentioning
confidence: 99%
“…The transformer architecture based on attention shows impressive results in many problems related to the processing of sequential data and in particular NLP [5], [9] and computer vision [8].…”
Section: Introductionmentioning
confidence: 99%
“…Since then, various adversarial attack methods have been proposed in computer vision (Goodfellow, Shlens, and Szegedy 2015;Carlini and Wagner 2017;. Adversarial attacks are not limited to the image domain; they are also possible in domains such as video classification (Wei et al 2019;Chen et al 2021b), text classification (Fursov et al 2021), and speaker recognition (Chen et al 2021a). One of the reasons underlying the current success of adversarial attacks in various domains and tasks is that adversarial attacks manipulate high-dimensional data.…”
Section: Introductionmentioning
confidence: 99%