2022
DOI: 10.48550/arxiv.2202.01711
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Algorithmic Fairness Datasets: the Story so Far

Alessandro Fabris,
Stefano Messina,
Gianmaria Silvello
et al.

Abstract: Data-driven algorithms are being studied and deployed in diverse domains to support critical decisions, directly impacting on people's wellbeing. As a result, a growing community of algorithmic fairness researchers has been investigating the equity of existing algorithms and proposing novel ones, advancing the understanding of the risks and opportunities of automated decision-making for different populations. Algorithmic fairness progress hinges on data, which can be used appropriately only if adequately docum… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 103 publications
(188 reference statements)
0
3
0
Order By: Relevance
“…The first three years of the conference were examined to understand the process by which categories emerge and become naturalized within a nascent community of practice. To target works primarily about algorithmic fairness, the author adapted the selection criteria proposed in Fabris et al 's [10] survey of data sets used in the algorithmic fairness literature and selected the subsample of these papers whose abstract contains at least one of the following strings, where the asterisk represents the wildcard character: *fair* (targeting, for example "fairness", "unfair"), *bias* ("biased", "debias"), *discriminat* ("antidiscrimination", "discriminatory"), disparate, *parit* ("parity", "disparities"). From this subsample, only the papers that deal directly with race by restricting to papers which contain at least one of the following strings race, *racism ("racism", "antiracism"), racial were selected.…”
Section: Samplementioning
confidence: 99%
“…The first three years of the conference were examined to understand the process by which categories emerge and become naturalized within a nascent community of practice. To target works primarily about algorithmic fairness, the author adapted the selection criteria proposed in Fabris et al 's [10] survey of data sets used in the algorithmic fairness literature and selected the subsample of these papers whose abstract contains at least one of the following strings, where the asterisk represents the wildcard character: *fair* (targeting, for example "fairness", "unfair"), *bias* ("biased", "debias"), *discriminat* ("antidiscrimination", "discriminatory"), disparate, *parit* ("parity", "disparities"). From this subsample, only the papers that deal directly with race by restricting to papers which contain at least one of the following strings race, *racism ("racism", "antiracism"), racial were selected.…”
Section: Samplementioning
confidence: 99%
“…Indeed, they have been demonstrated to be a source of racial [31,32] or gender [33] discrimination. Moreover, well-known datasets such as CelebA [34], Open Images [35] or ImageNet [3] lack of diversity-as shown in [36] or [37]-have resulted in imbalanced samples. Thus, state-of-the-art algorithms are unable to yield uniform performance over all sub.…”
Section: Improperly Sampled Training Datamentioning
confidence: 99%
“…Most existing fairness research focuses on group fairness, and it does so in the context of data types other than geospatial [13]. Individual fairness enforcement requires the evaluation of đť‘‚ (đť‘š 2 ) hard constraints, growing quadratically with the number of datapoints, which is more difficult to achieve efficiently.…”
Section: Experimental Evaluationmentioning
confidence: 99%