2020
DOI: 10.1016/j.engappai.2020.103641
|View full text |Cite
|
Sign up to set email alerts
|

TextTricker: Loss-based and gradient-based adversarial attacks on text classification models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 25 publications
(8 citation statements)
references
References 15 publications
0
8
0
Order By: Relevance
“…Therefore, we propose a scoring mechanism to identify the important tokens in a document which have a high impact on the final ranking result. Following [26,55], we first calculate the gradient magnitude with respect to each input unit. Then, we sum up the score for each dimension in the embedding space as the token-level importance score.…”
Section: Token Importance Rankingmentioning
confidence: 99%
“…Therefore, we propose a scoring mechanism to identify the important tokens in a document which have a high impact on the final ranking result. Following [26,55], we first calculate the gradient magnitude with respect to each input unit. Then, we sum up the score for each dimension in the embedding space as the token-level importance score.…”
Section: Token Importance Rankingmentioning
confidence: 99%
“…In NLP, Current gradient-based methods do not generate words directly through continuous perturbations, but use first-order approximation to enumerate substitution words (Cheng et al, 2019;Behjati et al, 2019;Xu and Du, 2020). This one-off approach may result in large step size perturbation, violating the hypothesis of local linearization.…”
Section: Perturbationmentioning
confidence: 99%
“…And the methods that directly employ the gradients to craft adversarial samples is not applicable in NLP. Current practices of textual adversarial attacks that employ first-order approximation to find substitute words are less effective for one-off searching and can violate the local linearization assumption (Cheng et al, 2019;Behjati et al, 2019;Xu and Du, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…The most well-known robustness attack methods include Fast Gradient Sign [31], Jacobian-based saliency map approach [34], C&W attack [7], etc. In recent years a great number of works like [19,9,59,37,18,62] have developed adversarial attack with a variety of methods and techniques.…”
Section: Adversarial Attacksmentioning
confidence: 99%