Proceedings of the 44th International Conference on Software Engineering 2022
DOI: 10.1145/3510003.3510091
|View full text |Cite
|
Sign up to set email alerts
|

Training data debugging for the fairness of machine learning software

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 27 publications
(13 citation statements)
references
References 30 publications
0
13
0
Order By: Relevance
“…Therefore, data determine the decision logic of ML software to a large extent [17], and data bias is considered a main root cause of ML software bias [48]. Data testing aims to detect different types of data bias, including checking whether the labels of training data are biased (label bias) [35], whether the distribution of training data implies an unexpected correlation between the sensitive attribute and the outcome label (selection bias) [49], whether the features of training data contain bias (feature bias) [50].…”
Section: Fairness Testing Componentsmentioning
confidence: 99%
See 4 more Smart Citations
“…Therefore, data determine the decision logic of ML software to a large extent [17], and data bias is considered a main root cause of ML software bias [48]. Data testing aims to detect different types of data bias, including checking whether the labels of training data are biased (label bias) [35], whether the distribution of training data implies an unexpected correlation between the sensitive attribute and the outcome label (selection bias) [49], whether the features of training data contain bias (feature bias) [50].…”
Section: Fairness Testing Componentsmentioning
confidence: 99%
“…For example, for demographic parity, researchers calculate the favorable rate among different demographic groups and detect fairness violations by comparing these rates. If the rate difference, called Statistical Parity Difference (SPD) in the software fairness literature [35], [38], [48], [50], [118], is beyond a threshold, the software under test is identified as containing fairness bugs.…”
Section: Statistical Measurements As Test Oraclesmentioning
confidence: 99%
See 3 more Smart Citations