2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.00113
|View full text |Cite
|
Sign up to set email alerts
|

Interpreting Attributions and Interactions of Adversarial Attacks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(11 citation statements)
references
References 19 publications
0
11
0
Order By: Relevance
“…The meta gradient adversarial attack (MGAA) method [227] utilized the meta learning to learn a generalized meta gradient by treating the attack against one model as one individual task, such that the meta gradient can be quickly fine-tuned to find the effective adversarial perturbations for new models. [187] empirically verified that "the adversarial transferability and the interactions inside adversarial perturbations are negatively correlated", and proposed an interaction loss to generate high transferable perturbations. In addition to above loss functions defined based on intermediate layer features, the reverse adversarial perturbation (RAP) attack [152] proposed a novel min-max loss function, where the adversarial example was perturbed by maximizing the adversarial loss, i.e., adding a reverse adversarial perturbation.…”
Section: Model-level Adversarial Transferabilitymentioning
confidence: 90%
“…The meta gradient adversarial attack (MGAA) method [227] utilized the meta learning to learn a generalized meta gradient by treating the attack against one model as one individual task, such that the meta gradient can be quickly fine-tuned to find the effective adversarial perturbations for new models. [187] empirically verified that "the adversarial transferability and the interactions inside adversarial perturbations are negatively correlated", and proposed an interaction loss to generate high transferable perturbations. In addition to above loss functions defined based on intermediate layer features, the reverse adversarial perturbation (RAP) attack [152] proposed a novel min-max loss function, where the adversarial example was perturbed by maximizing the adversarial loss, i.e., adding a reverse adversarial perturbation.…”
Section: Model-level Adversarial Transferabilitymentioning
confidence: 90%
“…Such dilemma motivates research on explanation techniques for DL models [40,76], aiming to explain DL models' decisions [5] and understand adversarial attacks [15,64] as well as defenses [75], thereby paving the way for building secure and trustworthy models. Explanation methods can be categorized as global explanation and local explanation in terms of the analysis object [12].…”
Section: Background 21 Explanation On Dnnmentioning
confidence: 99%
“…All the backdoors can achieve a high success rate in our evaluation. Explanations can be used in a wide range of applications, which include but are not limited to explaining model decisions [14], understanding adversarial attacks [64] and defenses [50], etc. Further, by assessing faithfulness, consistency between explanation methods, models, and humans can be achieved.…”
Section: Limitations and Benefitsmentioning
confidence: 99%
See 1 more Smart Citation
“…However, Szegedy et al [46] found that DNNs are vulnerable to adversarial examples, i.e., the maliciously crafted inputs that are indistinguishable from the correctly classified images but can induce misclassification on the target model. Such vulnerability poses significant threats when applying DNNs to security-critical applications, which also attracts broad attention to the security of DNNs [10,14,49,55,64,65].…”
Section: Introductionmentioning
confidence: 99%