2018
DOI: 10.1007/978-3-030-01231-1_31
|View full text |Cite
|
Sign up to set email alerts
|

ConvNets and ImageNet Beyond Accuracy: Understanding Mistakes and Uncovering Biases

Abstract: ConvNets and ImageNet have driven the recent success of deep learning for image classification. However, the marked slowdown in performance improvement combined with the lack of robustness of neural networks to adversarial examples and their tendency to exhibit undesirable biases question the reliability of these methods. This work investigates these questions from the perspective of the end-user by using human subject studies and explanations. The contribution of this study is threefold. We first experimental… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
97
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 122 publications
(100 citation statements)
references
References 29 publications
2
97
0
1
Order By: Relevance
“…This is in part due to the end-to-end nature of their training, which encourages models to exploit biased features if this leads to accurate classification. Prior works have mostly focused on uncovering and addressing different instances of bias in learned models, including gender bias [3,39,1] and racial bias [29]. However, the bias of the data itself has received less attention from the community.…”
Section: Related Workmentioning
confidence: 99%
“…This is in part due to the end-to-end nature of their training, which encourages models to exploit biased features if this leads to accurate classification. Prior works have mostly focused on uncovering and addressing different instances of bias in learned models, including gender bias [3,39,1] and racial bias [29]. However, the bias of the data itself has received less attention from the community.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, Stock and Cisse [49] presented empirical evidence that the performance of state-of-the-art classifiers on ImageNet [47] is largely underestimated -much of the reamining error is due to the fact that ImageNet's singlelabel annotation ignores the intrinsic multi-label nature of the images. Unlike ImageNet, multi-label datasets (e.g.…”
Section: Introductionmentioning
confidence: 99%
“…In this paper, we have surfaced various oracle issues in ILSVRC2012. That errors and ambiguity exist in ImageNet data has been recognized for long [17,22,23], but as an academic benchmark focused on object classification, the presence of the right class in the top-5 has in the past years been seen as sufficient to consider the classification task solved. However, the oracles we currently have at hand, and the representation and evaluation frameworks we currently employ within ML, may be insufficient when considering the deployment of ML components in real-world application scenarios.…”
Section: Resultsmentioning
confidence: 99%
“…Beyond this, at a semantic and more conceptual level, human judgment of ground truth may have been partially erroneous, potentially inconsistent, or even ethically undesired to become part of an automated decision pipeline, e.g., because it may encode biased or offensive assessments [7,23,24,27]. Such issues raise questions on the true quality of the oracle, will have a serious influence on what an ML system will infer, and, from an experimental validity perspective, may cause systems to not actually learn what they are supposed to learn [24].…”
mentioning
confidence: 99%