Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.488
|View full text |Cite
|
Sign up to set email alerts
|

Towards Debiasing Sentence Representations

Abstract: As natural language processing methods are increasingly deployed in real-world scenarios such as healthcare, legal systems, and social science, it becomes necessary to recognize the role they potentially play in shaping social biases and stereotypes. Previous work has revealed the presence of social biases in widely used word embeddings involving gender, race, religion, and other social constructs. While some methods were proposed to debias these word-level embeddings, there is a need to perform debiasing at t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
61
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 85 publications
(89 citation statements)
references
References 46 publications
1
61
0
Order By: Relevance
“…Despite the relatively ample literature on gender debiasing for word-level representations, very little work has focused on sentence representations (Liang et al, 2020;Liu et al, 2019;Lee et al, 2019b). Until this point, most debiasing work on sentences mainly focus on measuring bias (Lee et al, 2019b;.…”
Section: Related Workmentioning
confidence: 99%
“…Despite the relatively ample literature on gender debiasing for word-level representations, very little work has focused on sentence representations (Liang et al, 2020;Liu et al, 2019;Lee et al, 2019b). Until this point, most debiasing work on sentences mainly focus on measuring bias (Lee et al, 2019b;.…”
Section: Related Workmentioning
confidence: 99%
“…To sum up, people reveal their stereotype expectancies in many subtle ways in the words they use. This fact can explain the effectiveness of several computational works at measuring social bias (e.g., gender, racial, religion, and ethnic stereotypes among others) by using word representations [28,41,42]. In [41], social bias are quantified by using embeddings of representative words such as women, men, Asians living in the United States, and white people (i.e., non-Hispanic subpopulation from the United States).…”
Section: Related Workmentioning
confidence: 99%
“…In [28,42], word embeddings are used to measure (and reduce) the bias. In [28], a methodology based directly on word embeddings is proposed to differentiate gender bias associations (e.g., biased association between receptionist and female) from associations of related concepts (e.g., between queen and female); a neutralisation and equalisation debias process.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In four tasks with known bias factors -hate speech detection, toxicity detection, occupation prediction from short bios, and coreference resolution -we explore whether upstream bias mitigation of a LM followed by downstream fine-tuning reduces bias for the downstream model. Though previous work has addressed biases in frozen PTLM or word embeddings (Bolukbasi et al, 2016;Zhou et al, 2019;Bhardwaj et al, 2020;Liang et al, 2020;Ravfogel et al, 2020), for example by measuring associations between gender and occupations in an embedding space, they do not study their effect on downstream classifiers (Fig. 1 (b)), while some of them study the effects while keeping the embeddings frozen Kurita et al, 2019;Prost et al, 2019).…”
Section: Introductionmentioning
confidence: 99%