2022
DOI: 10.1007/978-3-031-20065-6_18
|View full text |Cite
|
Sign up to set email alerts
|

Scaling Adversarial Training to Large Perturbation Bounds

Abstract: The vulnerability of Deep Neural Networks to Adversarial Attacks has fuelled research towards building robust models. While most Adversarial Training algorithms aim at defending attacks constrained within low magnitude Lp norm bounds, real-world adversaries are not limited by such constraints. In this work, we aim to achieve adversarial robustness within larger bounds, against perturbations that may be perceptible, but do not change human (or Oracle) prediction. The presence of images that flip Oracle predicti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(7 citation statements)
references
References 19 publications
0
7
0
Order By: Relevance
“…in which x is the adversarial data generated via PGD within the -ball centered at x and d(•, •) : V × V → R is a distance function, such as the Kullback-Leibler (KL) divergence , the Jensen-Shannon (JS) divergence (Addepalli et al, 2022), and the optimal transport (OT) distance (Zhang and Wang, 2019). We denote the RD on the unlabeled set X as L RD (X; θ) = xi∈X RD (x i ; θ).…”
Section: Representational Divergence (Rd)mentioning
confidence: 99%
See 1 more Smart Citation
“…in which x is the adversarial data generated via PGD within the -ball centered at x and d(•, •) : V × V → R is a distance function, such as the Kullback-Leibler (KL) divergence , the Jensen-Shannon (JS) divergence (Addepalli et al, 2022), and the optimal transport (OT) distance (Zhang and Wang, 2019). We denote the RD on the unlabeled set X as L RD (X; θ) = xi∈X RD (x i ; θ).…”
Section: Representational Divergence (Rd)mentioning
confidence: 99%
“…This section provides the results of ACL with RCS using different distance functions including the KL divergence , the JS divergence (Addepalli et al, 2022), and the OT distance (Zhang and Wang, 2019) for calculating the RD L RD (•). Other training settings exactly keep the same as Section 4.1.…”
Section: B5 Efficient Acl Via Rcs With Various Distance Functions φmentioning
confidence: 99%
“…For the term g(w c , (k) , t = t) 2−q to become o(1) at the time of convergence, 1+o (1) 2n tσ 2 − (q − 2)σ 2−q 0 σ 2−q should be constant. Equating the L.H.S.…”
Section: Convergence Time Of Noisy Patchesmentioning
confidence: 99%
“…Using this, we obtain a strong benchmark for ID generalization as shown in Table-1. However, as shown in prior works [1], the impact of augmentations in training is limited by the capacity of the network in being able to generalize well to the diverse augmented data distribution. Therefore, increasing the diversity of training data demands the use of larger model capacities to achieve optimal performance.…”
Section: Introductionmentioning
confidence: 99%
“…Adversarial training serves as the foundation for various defense methods, including those employing strong data augmentation [31], auxiliary data for primary task robustness [32], and class-fairness considerations [33]. Despite the success of these adversarial defense methods, they mainly focus on the single-mode setting while ignoring the fact that real-world datasets usually have large intra-variations or multiple modes depending on data labeling.…”
Section: Introductionmentioning
confidence: 99%