Findings of the Association for Computational Linguistics: ACL 2022 2022
DOI: 10.18653/v1/2022.findings-acl.289
|View full text |Cite
|
Sign up to set email alerts
|

Detection of Adversarial Examples in Text Classification: Benchmark and Baseline via Robust Density Estimation

Abstract: Word-level adversarial attacks have shown success in NLP models, drastically decreasing the performance of transformer-based models in recent years. As a countermeasure, adversarial defense has been explored, but relatively few efforts have been made to detect adversarial examples. However, detecting adversarial examples may be crucial for automated tasks (e.g. review sentiment analysis) that wish to amass information about a certain population and additionally be a step towards a robust defense system. To thi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 13 publications
0
7
0
Order By: Relevance
“…DISP performs very well on AGNEWS, which may be due to the synonyms replaced by these attack algorithms do not preserve the semantics of the original sentences well. 2) Consistently with Yoo et al (2022), FGWS works badly in the face of more subtle attacks, such as BAE and TextFooler. 3) Both RDE and MD are feature density-based methods, and in general, RDE works better than MD.…”
Section: Detection Performancementioning
confidence: 95%
See 3 more Smart Citations
“…DISP performs very well on AGNEWS, which may be due to the synonyms replaced by these attack algorithms do not preserve the semantics of the original sentences well. 2) Consistently with Yoo et al (2022), FGWS works badly in the face of more subtle attacks, such as BAE and TextFooler. 3) Both RDE and MD are feature density-based methods, and in general, RDE works better than MD.…”
Section: Detection Performancementioning
confidence: 95%
“…Following the work of Yoo et al (2022), we divide the detection of adversarial samples into two scenarios. Scenario 1 will detect all adversarial samples, regardless of whether the model output is successfully changed or not.…”
Section: Detection Performancementioning
confidence: 99%
See 2 more Smart Citations
“…Enhancing reliability can be accomplished through the use of advanced uncertainty estimation (UE) techniques (Lakshminarayanan et al, 2017;Gal and Ghahramani, 2016;Lee et al, 2018;Liu et al, 2020;Podolskiy et al, 2021;Xin et al, 2021;Yoo et al, 2022). Promoting model fairness entails defining fairness metrics and employing special debiasing techniques (Elazar and Goldberg, 2018;Wang et al, 2019;Ravfogel et al, 2020;Han et al, 2021Han et al, , 2022aBaldini et al, 2022).…”
Section: Introductionmentioning
confidence: 99%