2022
DOI: 10.48550/arxiv.2203.08392
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Patch-Fool: Are Vision Transformers Always Robust Against Adversarial Perturbations?

Abstract: Vision transformers (ViTs) have recently set off a new wave in neural architecture design thanks to their record-breaking performance in various vision tasks. In parallel, to fulfill the goal of deploying ViTs into real-world vision applications, their robustness against potential malicious attacks has gained increasing attention. In particular, recent works show that ViTs are more robust against adversarial attacks as compared with convolutional neural networks (CNNs), and conjecture that this is because ViTs… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(13 citation statements)
references
References 44 publications
0
13
0
Order By: Relevance
“…After obtaining the bias-only model, the following procedure in Step 2 of the DSA framework localizes and masks the spurious (sensitive) features via adversarial attacks that are generated using the Patch-Fool construction proposed in [13]. Specifically, Patch-Fool is designed to fool the self-attention mechanism in ViTs by attacking their basic component (i.e., a single patch) with a series of attentionaware optimization techniques, demonstrating that the ViTs are more vulnerable to adversarial attacks than the CNNs.…”
Section: Adversarial Attack Against the Bias-only Modelmentioning
confidence: 99%
See 4 more Smart Citations
“…After obtaining the bias-only model, the following procedure in Step 2 of the DSA framework localizes and masks the spurious (sensitive) features via adversarial attacks that are generated using the Patch-Fool construction proposed in [13]. Specifically, Patch-Fool is designed to fool the self-attention mechanism in ViTs by attacking their basic component (i.e., a single patch) with a series of attentionaware optimization techniques, demonstrating that the ViTs are more vulnerable to adversarial attacks than the CNNs.…”
Section: Adversarial Attack Against the Bias-only Modelmentioning
confidence: 99%
“…Specifically, Patch-Fool is designed to fool the self-attention mechanism in ViTs by attacking their basic component (i.e., a single patch) with a series of attentionaware optimization techniques, demonstrating that the ViTs are more vulnerable to adversarial attacks than the CNNs. However, in contrast to [13], instead of applying Patch-Fool as an adversarial attack method to evaluate the robustness of ViT, we utilize it to efficiently localize and mask the sensitive features in the inputs. To this end, we adapt the objective function of Patch-Fool in order to attack the bias-only model on the sensitive labels instead of the target labels.…”
Section: Adversarial Attack Against the Bias-only Modelmentioning
confidence: 99%
See 3 more Smart Citations