Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-main.188
|View full text |Cite
|
Sign up to set email alerts
|

How Gender Debiasing Affects Internal Model Representations, and Why It Matters

Abstract: Common studies of gender bias in NLP focus either on extrinsic bias measured by model performance on a downstream task or on intrinsic bias found in models' internal representations. However, the relationship between extrinsic and intrinsic bias is relatively unknown. In this work, we illuminate this relationship by measuring both quantities together: we debias a model during downstream fine-tuning, which reduces extrinsic bias, and measure the effect on intrinsic bias, which is operationalized as bias extract… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(22 citation statements)
references
References 30 publications
1
21
0
Order By: Relevance
“…Our case study extends the experiments done by Orgad et al (2022). In their work, they tested a RoBERTa-based classifier finetuned on Bias in Bios.…”
Section: Case Study: Balancing the Test Datamentioning
confidence: 77%
See 4 more Smart Citations
“…Our case study extends the experiments done by Orgad et al (2022). In their work, they tested a RoBERTa-based classifier finetuned on Bias in Bios.…”
Section: Case Study: Balancing the Test Datamentioning
confidence: 77%
“…Winobias sentences consist of an anti-and a pro-stereotypical sentence, as shown in Figure 1. Coreference systems should be able to resolve both sentences correctly, but most perform poorly on the anti-stereotypical ones Orgad et al, 2022).…”
Section: Case Study: Winobiasmentioning
confidence: 99%
See 3 more Smart Citations