2020
DOI: 10.1109/access.2020.3024244
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Class Triplet Loss With Gaussian Noise for Adversarial Robustness

Abstract: Deep Neural Networks (DNNs) classifiers performance degrades under adversarial attacks, such attacks are indistinguishably perturbed relative to the original data. Providing robustness to adversarial attacks is an important challenge in DNN training, which has led to extensive research. In this paper, we harden DNN classifiers under the adversarial attacks by regularizing their deep internal representation space with Multi-class Triplet regularization method. This method enables DNN classifier to learn a featu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 22 publications
0
1
0
Order By: Relevance
“…Adversarial attacks have been investigated in the area of image, audios, texts and recently in Windows executable files classification exercise and a number of successful attacks examples in image, audios and texts have been generated to cause misclassification [1][2][3][4][5][6][7]. The principal reason for the success in image, audios and texts is that their feature-space is comparatively fixed, an image or text can be formatted as a three-dimensional array of pixels with each pixel value as a three-dimensional RGB (red, green, blue) vector value ranged from 0 to 255, thus, is feasible to find an exact function that is differentiable, therefore, a feature-space attack built on gradients can instantly apply on text or images to create adversarial attack examples.…”
Section: Introductionmentioning
confidence: 99%
“…Adversarial attacks have been investigated in the area of image, audios, texts and recently in Windows executable files classification exercise and a number of successful attacks examples in image, audios and texts have been generated to cause misclassification [1][2][3][4][5][6][7]. The principal reason for the success in image, audios and texts is that their feature-space is comparatively fixed, an image or text can be formatted as a three-dimensional array of pixels with each pixel value as a three-dimensional RGB (red, green, blue) vector value ranged from 0 to 255, thus, is feasible to find an exact function that is differentiable, therefore, a feature-space attack built on gradients can instantly apply on text or images to create adversarial attack examples.…”
Section: Introductionmentioning
confidence: 99%