2021
DOI: 10.48550/arxiv.2110.06177
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Tracking the risk of a deployed model and detecting harmful distribution shifts

Abstract: When deployed in the real world, machine learning models inevitably encounter changes in the data distribution, and certain-but not all-distribution shifts could result in significant performance degradation. In practice, it may make sense to ignore benign shifts, under which the performance of a deployed model does not degrade substantially, making interventions by a human expert (or model retraining) unnecessary. While several works have developed tests for distribution shifts, these typically either use non… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 9 publications
0
1
0
Order By: Relevance
“…Considering the problems with the accuracy metrics discussed in the previous section, it is also encouraging that there are various measures taken within the ML community to address issues. For example, continuous performance monitoring seeks to identify and fix deteriorating performance over time (Podkopaev and Ramdas 2022), uncertainty quantification provides insights into unknowns in testing (Siddique et al 2022), and methods for interpretability can aid in identifying spurious and problematic correlations that inflate accuracy numbers (Poechhacker and Kacianka 2021). In particular, external auditing is a promising quantitative approach for constructing different, partial accounts of opaque algorithms that often center perspectives and experiences of marginalized people to make possible harms visible (Sandvig et al 2014;Metaxa et al 2021).…”
Section: Rethinking Accuracymentioning
confidence: 99%
“…Considering the problems with the accuracy metrics discussed in the previous section, it is also encouraging that there are various measures taken within the ML community to address issues. For example, continuous performance monitoring seeks to identify and fix deteriorating performance over time (Podkopaev and Ramdas 2022), uncertainty quantification provides insights into unknowns in testing (Siddique et al 2022), and methods for interpretability can aid in identifying spurious and problematic correlations that inflate accuracy numbers (Poechhacker and Kacianka 2021). In particular, external auditing is a promising quantitative approach for constructing different, partial accounts of opaque algorithms that often center perspectives and experiences of marginalized people to make possible harms visible (Sandvig et al 2014;Metaxa et al 2021).…”
Section: Rethinking Accuracymentioning
confidence: 99%