2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.01482
|View full text |Cite
|
Sign up to set email alerts
|

Are Labels Always Necessary for Classifier Accuracy Evaluation?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
17
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 58 publications
(17 citation statements)
references
References 15 publications
0
17
0
Order By: Relevance
“…A separate line of work departs from complexity measures altogether and directly predicts OOD generalization from unlabelled test data. These methods either predict the correctness of the model directly on individual examples [14,32,15] or directly estimate the total error [19,24,9,10,68]. Although these methods work well in practice, they do not provide any insight into the underlying mechanism of generalization since they act only on the output layer of the network.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…A separate line of work departs from complexity measures altogether and directly predicts OOD generalization from unlabelled test data. These methods either predict the correctness of the model directly on individual examples [14,32,15] or directly estimate the total error [19,24,9,10,68]. Although these methods work well in practice, they do not provide any insight into the underlying mechanism of generalization since they act only on the output layer of the network.…”
Section: Related Workmentioning
confidence: 99%
“…Directly estimating the generalization of a trained model on test data is one approach to this problem [14,32,15,19]. However, these methods are typically calculated based on the output predictive distribution of a model, which can become poorly calibrated in out-of-domain settings.…”
Section: Introductionmentioning
confidence: 99%
“…This work has only recently been applied to information security, specifically for Windows malware classification [27]. Directly targeting model evaluation instead, Deng and Zheng [28] proposed AutoEval to estimate the accuracy of a classifier on an unlabeled dataset by using feature statistics from the training set and synthetic datasets generated by applying transformations to the training set. Novák et al [29] use density estimation to similar ends.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, several studies outside the field of TL have started to establish a connection between the distribution divergence and accuracy. Among them, the most related work is from Elsahar and Galle [11], [12], who used various methods such as H-divergence, Fréchet distance and confidence-based metrics to predict the accuracy drop of modern NLP and computer version (CV) models under domain shifts. Both studies above used predicted accuracy drops to evaluate the robustness of trained models.…”
Section: Related Workmentioning
confidence: 99%