Proceedings of the 29th ACM SIGSOFT International Symposium on Software Testing and Analysis 2020
DOI: 10.1145/3395363.3397357
|View full text |Cite
|
Sign up to set email alerts
|

DeepGini: prioritizing massive tests to enhance the robustness of deep neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
162
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 173 publications
(164 citation statements)
references
References 31 publications
2
162
0
Order By: Relevance
“…So far this problem remains an open challenge. Prior work like DeepGini has proposed to calculate a Gini index of a test case from the model's output probability distribution [7]. DeepGini's intuition is to favor those test cases with most uncertainty (e.g., a more flat distribution) under the current model's prediction.…”
Section: Fol Guided Test Case Selectionmentioning
confidence: 99%
See 1 more Smart Citation
“…So far this problem remains an open challenge. Prior work like DeepGini has proposed to calculate a Gini index of a test case from the model's output probability distribution [7]. DeepGini's intuition is to favor those test cases with most uncertainty (e.g., a more flat distribution) under the current model's prediction.…”
Section: Fol Guided Test Case Selectionmentioning
confidence: 99%
“…We adopt the most recent work DeepGini[7] as the baseline of the test case selection strategy. DeepGini calculates a Gini index for each test case according to the output probability distribution of the model.…”
mentioning
confidence: 99%
“…Kim et al [17] proposed surprise guided testing metrics based on the similarity between the training data and test data. Moreover, some prediction probability based data selection metrics [18], [19], [70] are also proposed. Most recently, Wang et al [71] proposed a robustness-oriented data selection metric, however, their metric can only select data that are generated by adversarial attacks, it is out of our consideration.…”
Section: Related Workmentioning
confidence: 99%
“…On the other hand, recent work on DL testing and debugging [12]- [19] have proposed different metrics for test generation and test selection, i.e., the problem of selecting test data that are more likely to be misclassified by the model [20]. As in active learning scenarios, these test data can then be used to improve the model (by retraining).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation