2019
DOI: 10.5120/ijca2019919384
|View full text |Cite
|
Sign up to set email alerts
|

Toward Mitigating Adversarial Texts

Abstract: Neural networks are frequently used for text classification, but can be vulnerable to misclassification caused by adversarial examples: input produced by introducing small perturbations that cause the neural network to output an incorrect classification. Previous attempts to generate black-box adversarial texts have included variations of generating nonword misspellings, natural noise, synthetic noise, along with lexical substitutions. This paper proposes a defense against black-box adversarial attacks using a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
4
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(5 citation statements)
references
References 17 publications
1
4
0
Order By: Relevance
“…(2) The Bi-LSTM model is more vulnerable to the two attacks than the CNN model by a 12.45% accuracy difference on average. This supports the conclusion from previous research that, in the NLP domain, deep CNNs tend to be more robust than RNN models (Ren et al, 2019;Alshemali and Kalita, 2019 Table 3: The accuracy of the nonneural classification models under adversarial attacks, with and without the defense applied. Percent Increase is the percent increase of the classification accuracy with the defense applied.…”
Section: Effectiveness Of the Defensesupporting
confidence: 88%
“…(2) The Bi-LSTM model is more vulnerable to the two attacks than the CNN model by a 12.45% accuracy difference on average. This supports the conclusion from previous research that, in the NLP domain, deep CNNs tend to be more robust than RNN models (Ren et al, 2019;Alshemali and Kalita, 2019 Table 3: The accuracy of the nonneural classification models under adversarial attacks, with and without the defense applied. Percent Increase is the percent increase of the classification accuracy with the defense applied.…”
Section: Effectiveness Of the Defensesupporting
confidence: 88%
“…Defenses based on spell and syntax checkers are successful against character-level text attacks (Pruthi et al, 2019;Wang et al, 2019;Alshemali and Kalita, 2019). In contrast, these solutions are not effective against word-level attacks preserving language correctness (Wang et al, 2019).…”
Section: Defense Against Adversarial Attacks In Nlpmentioning
confidence: 99%
“…Following the methodology used in [2] to test the spellcheckers, we generated four types of adversarial text including our attack using a list of top 20 frequent words in the SMS dataset. We used the DeepWordBug method proposed by Gao et al [17] to generate three adversarial texts: (1) insertion: inserted one random character to the words (e.g., c*all), (2) deletion: we removed the second character (one was removed per word), and (3) swapping: we swapped the second and third characters in the word (one swap per word).…”
Section: Effect Of Adversarial Text On Auto-correctionmentioning
confidence: 99%
“…On the other hand, using spell checking algorithms is the most common defence method against character-level perturbation in NLP tasks [29]. Although spell checking methods could detect and correct errors or adversarial examples, they cannot be applied in all domains because their performance varies depending on the type of misspelling [2].…”
Section: Introductionmentioning
confidence: 99%