Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society 2018
DOI: 10.1145/3278721.3278729
|View full text |Cite
|
Sign up to set email alerts
|

Measuring and Mitigating Unintended Bias in Text Classification

Abstract: We introduce and illustrate a new approach to measuring and mitigating unintended bias in machine learning models. Our definition of unintended bias is parameterized by a test set and a subset of input features. We illustrate how this can be used to evaluate text classifiers using a synthetic test set and a public corpus of comments annotated for toxicity from Wikipedia Talk pages. We also demonstrate how imbalances in training data can lead to unintended bias in the resulting models, and therefore potentially… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

11
670
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 552 publications
(681 citation statements)
references
References 5 publications
11
670
0
Order By: Relevance
“…Results are shown in Table 4. Unlike previous approaches (Park et al, 2018;Dixon et al, 2018;Madras et al, 2018), our method does not degrade classifier performance (it even improves) in terms of all reported metrics. We also look at samples containing identity terms.…”
Section: Evaluation On Original Datamentioning
confidence: 48%
“…Results are shown in Table 4. Unlike previous approaches (Park et al, 2018;Dixon et al, 2018;Madras et al, 2018), our method does not degrade classifier performance (it even improves) in terms of all reported metrics. We also look at samples containing identity terms.…”
Section: Evaluation On Original Datamentioning
confidence: 48%
“…"Wiki Madlibs" Eval Dataset [8]: For the Dataset augmentation based bias mitigation baseline strategy, we follow the approach proposed by Dixon et al [8]. For every BSW, they manually create templates for both hateful instances and neutral instances.…”
Section: Datasetsmentioning
confidence: 99%
“…In order to empirically measure the performance of the de-biasing procedures, Dixon et al [8] proposed the Pinned AUC Equality Difference (pAUC) metric to quantify the bias in learning for the dataset augmentation strategy only.…”
Section: Pinned Auc Equality Differencementioning
confidence: 99%
See 1 more Smart Citation
“…For the overall performance of these models, we will use the standard accuracy metric of multi-class classification. There are a number of metrics and criteria that offer different interpretations of model fairness (Dixon et al, 2018;Narayanan, 2018;Friedler et al, 2019;. In this work, we use the method introduced by (Hardt et al, 2016) as Equality of Opportunity.…”
Section: Evaluation Metricsmentioning
confidence: 99%