2019
DOI: 10.48550/arxiv.1903.12261
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Benchmarking Neural Network Robustness to Common Corruptions and Perturbations

Abstract: In this paper we establish rigorous benchmarks for image classifier robustness. Our first benchmark, IMAGENET-C, standardizes and expands the corruption robustness topic, while showing which classifiers are preferable in safety-critical applications. Then we propose a new dataset called IMAGENET-P which enables researchers to benchmark a classifier's robustness to common perturbations. Unlike recent robustness research, this benchmark evaluates performance on common corruptions and perturbations not worst-case… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
481
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 281 publications
(481 citation statements)
references
References 23 publications
0
481
0
Order By: Relevance
“…The robustness and out-of-distribution (OOD) generalization abilities of RELICv2 representations are tested on several detasets. We use ImageNetV2 (Recht et al, 2019) and ImageNet-C (Hendrycks and Dietterich, 2019) datasets to evaluate robustness. ImageNetV2 (Recht et al, 2019) has three sets of 10000 images that were collected to have a similar distribution to the original ImageNet validation set, while ImageNet-C (Hendrycks and Dietterich, 2019) consists of 15 synthetically generated corruptions (e.g.…”
Section: B5 Robustness and Ood Generalizationmentioning
confidence: 99%
See 1 more Smart Citation
“…The robustness and out-of-distribution (OOD) generalization abilities of RELICv2 representations are tested on several detasets. We use ImageNetV2 (Recht et al, 2019) and ImageNet-C (Hendrycks and Dietterich, 2019) datasets to evaluate robustness. ImageNetV2 (Recht et al, 2019) has three sets of 10000 images that were collected to have a similar distribution to the original ImageNet validation set, while ImageNet-C (Hendrycks and Dietterich, 2019) consists of 15 synthetically generated corruptions (e.g.…”
Section: B5 Robustness and Ood Generalizationmentioning
confidence: 99%
“…We use ImageNetV2 (Recht et al, 2019) and ImageNet-C (Hendrycks and Dietterich, 2019) datasets to evaluate robustness. ImageNetV2 (Recht et al, 2019) has three sets of 10000 images that were collected to have a similar distribution to the original ImageNet validation set, while ImageNet-C (Hendrycks and Dietterich, 2019) consists of 15 synthetically generated corruptions (e.g. blur, noise) that are added to the ImageNet validation set.…”
Section: B5 Robustness and Ood Generalizationmentioning
confidence: 99%
“…Deep neural networks are known to be vulnerable to adversarial examples and common corruptions (Bulusu et al, 2020). Hendrycks & Dietterich (2019); Hendrycks et al (2021) developed corruption robustness benchmarking datasets CIFAR-10/100-C, ImageNet-C, and ImageNet-R to facilitate robustness evaluations of CIFAR and ImageNet classification models. Michaelis et al (2019) extended this benchmark to object detection models.…”
Section: Related Workmentioning
confidence: 99%
“…To bridge this gap, we design 15 common corruptions for benchmarking corruption robustness of point cloud recognition models. It is worth noting that such designs are non-trivial since the manipulation space of 3D point clouds is completely different from 2D images where the corruptions come from the RGB modification (Hendrycks & Dietterich, 2019). In particular, we have three principles to design our benchmarks: i) Since we directly manipulate the position of points, we need to take extra care to preserve the original semantics of point clouds (Fig.…”
Section: D Point Cloud Corruption Robustnessmentioning
confidence: 99%
See 1 more Smart Citation