Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.446
|View full text |Cite
|
Sign up to set email alerts
|

Monolingual and Multilingual Reduction of Gender Bias in Contextualized Representations

Abstract: Pretrained language models (PLMs) learn stereotypes held by humans and reflected in text from their training corpora, including gender bias. When PLMs are used for downstream tasks such as picking candidates for a job, people's lives can be negatively affected by these learned stereotypes. Prior work usually identifies a linear gender subspace and removes gender information by eliminating the subspace. Following this line of work, we propose to use DensRay, an analytical method for obtaining interpretable dens… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
27
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 27 publications
(27 citation statements)
references
References 21 publications
0
27
0
Order By: Relevance
“…Considering inprocessing, we may re-train the models with additional fairness constraints or introduce adversarial training [1128,1144,1130,1145]. For post-processing, transfer learning [1146] is an option, e.g., fine-tuned language models on English to address its efficacy on Chinese as well.…”
Section: Fairnessmentioning
confidence: 99%
“…Considering inprocessing, we may re-train the models with additional fairness constraints or introduce adversarial training [1128,1144,1130,1145]. For post-processing, transfer learning [1146] is an option, e.g., fine-tuned language models on English to address its efficacy on Chinese as well.…”
Section: Fairnessmentioning
confidence: 99%
“…Transfer learning is another option. Liang et al (2020c) makes use of fine-tuned multilingual LM on English to address its efficacy on Chinese as well. reveals the presence of gender bias and proposes a method to mitigate in multilingual word embeddings using alignment.…”
Section: Related Workmentioning
confidence: 99%
“…In the following, we focus on approaches for mitigating biases from PLMs, which are largely inspired by debiasing for static word embeddings (e.g., Bolukbasi et al, 2016;Dev and Phillips, 2019;Lauscher et al, 2020a;Karve et al, 2019, inter alia). While several works propose projection-based debiasing for PLMs (e.g., Dev et al, 2020;Liang et al, 2020;Kaneko and Bollegala, 2021), most of the debiasing approaches require training. Here, some methods rely on debiasing objectives (e.g., Qian et al, 2019;Bordia and Bowman, 2019).…”
Section: Related Workmentioning
confidence: 99%