2023
DOI: 10.1016/j.patcog.2023.109308
|View full text |Cite
|
Sign up to set email alerts
|

A comprehensive evaluation framework for deep model robustness

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 24 publications
(10 citation statements)
references
References 6 publications
0
2
0
Order By: Relevance
“…Several physical-world adversarial example generation methods have been proposed and demonstrated to be effective [10,13,[34][35][36][37][38][39]. However, they use different datasets for evaluation, which makes it difficult to conduct a comprehensive evaluation.…”
Section: Adversarial Robustness Benchmarkmentioning
confidence: 99%
See 1 more Smart Citation
“…Several physical-world adversarial example generation methods have been proposed and demonstrated to be effective [10,13,[34][35][36][37][38][39]. However, they use different datasets for evaluation, which makes it difficult to conduct a comprehensive evaluation.…”
Section: Adversarial Robustness Benchmarkmentioning
confidence: 99%
“…Tang [42] proposed the first unified Robustness Assessment Benchmark, RobustART, which provides a standardized evaluation framework for adversarial examples. Recently, there has been a notable emergence of a series of benchmarks [39,[43][44][45][46][47][48] in different fields. These benchmarks serve the crucial function of unifying evaluation metrics, enabling a clear and systematic comparison of the strengths and weaknesses inherent in various methods.…”
Section: Adversarial Robustness Benchmarkmentioning
confidence: 99%
“…In order to comprehensively evaluate the security of deep learning models, some research [14][15][16][17][18][19][20][21][22] has proposed evaluation metrics, method libraries, and evaluation platforms. Works such as CleverHans [14] and FoolBox [15] integrated the most common attack and defense methods, but the neglect of code quality made some of them incorrectly developed 2 of 43 or even unworkable.…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, multiple objectives could be at odds with each other [21,47,54] and selecting features satisfying more than one objective still remains non-trivial. As research in responsible AI and regulatory requirements for machine learning models rapidly advance, new metrics are being developed to evaluate model performance, both within [28,41] and beyond [49] the secondary characteristics discussed Hence, we introduce the process of feature reselection. Feature reselection aims to select features to improve on secondary model performance characteristics (characteristics that become important to consider after a model is already developed) while maintaining similar performance with respect to a primary characteristic based on which features were already selected.…”
Section: Introductionmentioning
confidence: 99%