2022
DOI: 10.48550/arxiv.2202.03277
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On The Empirical Effectiveness of Unrealistic Adversarial Hardening Against Realistic Adversarial Attacks

Abstract: While the literature on security attacks and defense of Machine Learning (ML) systems mostly focuses on unrealistic adversarial examples, recent research has raised concern about the under-explored field of realistic adversarial attacks and their implications on the robustness of real-world systems. Our paper paves the way for a better understanding of adversarial robustness against realistic attacks and makes two major contributions. First, we conduct a study on three real-world use cases (text classification… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 39 publications
0
2
0
Order By: Relevance
“…Finally, because the perturbation size is known to impact success rate and human perceptibility of adversarial attacks in other domains (Simonetto et al, 2021;Dyrmishi et al, 2022), we investigate the relationship between the number of altered words and validity/naturalness.…”
Section: Rq2 (Naturalness): Are Adversarial Examplesmentioning
confidence: 99%
“…Finally, because the perturbation size is known to impact success rate and human perceptibility of adversarial attacks in other domains (Simonetto et al, 2021;Dyrmishi et al, 2022), we investigate the relationship between the number of altered words and validity/naturalness.…”
Section: Rq2 (Naturalness): Are Adversarial Examplesmentioning
confidence: 99%
“…where θ are the parameters of the network, x is the input data, y i is its associated target, L(θ, x, y i ) is the loss function used, and the strength of the attack. Following Goodfellow, other attacks were proposed, first by adding iterations (I-FGSM) (Kurakin, Goodfellow, and Bengio 2016), projections and random restart (PGD) (Madry et al 2017), momentum (MIM) (Dong et al 2018) and constraints (CoEva2) (Ghamizi et al 2020;Dyrmishi et al 2022). These algorithms can be used without any change on a multi-task model if the attacker only focuses on a single task.…”
Section: Related Workmentioning
confidence: 99%