2021
DOI: 10.48550/arxiv.2106.05498
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

It's COMPASlicated: The Messy Relationship between RAI Datasets and Algorithmic Fairness Benchmarks

Abstract: Risk assessment instrument (RAI) datasets, particularly ProPublica's COMPAS dataset, are commonly used in algorithmic fairness papers due to benchmarking practices of comparing algorithms on datasets used in prior work. In many cases, this data is used as a benchmark to demonstrate good performance without accounting for the complexities of criminal justice (CJ) processes. We show that pretrial RAI datasets contain numerous measurement biases and errors inherent to CJ pretrial evidence and due to disparities i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(28 citation statements)
references
References 71 publications
0
28
0
Order By: Relevance
“…Indeed, it is now accepted that race is a social construct and that there is greater genetic variability within a particular race than there is between races [157][158][159] . As such, categorization of patients by race can obscure a host of potential confounders to fairness analyses including culture, history, and socioeconomic status that all may separately and synergistically influence a particular patient's health 160,161 . These manifold factors can also vary by location so that the same person may be considered of different races in different geographic locations, as seen in the example of self-reported Asian ethnicity in the TCGA versus Pioneer and self-reported race in COMPAS 55,161 .…”
Section: Fragility Of Racementioning
confidence: 99%
“…Indeed, it is now accepted that race is a social construct and that there is greater genetic variability within a particular race than there is between races [157][158][159] . As such, categorization of patients by race can obscure a host of potential confounders to fairness analyses including culture, history, and socioeconomic status that all may separately and synergistically influence a particular patient's health 160,161 . These manifold factors can also vary by location so that the same person may be considered of different races in different geographic locations, as seen in the example of self-reported Asian ethnicity in the TCGA versus Pioneer and self-reported race in COMPAS 55,161 .…”
Section: Fragility Of Racementioning
confidence: 99%
“…We refer to the combination of these two properties throughout the paper as smooth general capability (or performance) scaling. 6 In Section 2.2, we discuss abrupt specific capability scaling, in which models can also suddenly gain specific capabilities at scale. We illustrate this phenomenon with three examples from the literature [11,56,4].…”
Section: Introductionmentioning
confidence: 99%
“…As far as datasets go, we have conducted experiments on 4 datasets that have been used extensively in the fairness literature to benchmark different fairness enhancing interventions. Recent research efforts have questioned the validity of some these datasets like the COMPAS [3] and Adult Income dataset [12]. Future work might include more recent datasets [12] and other stages of intervention, such as the data curation stage into the analysis [22].…”
Section: Discussionmentioning
confidence: 99%