Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency 2020
DOI: 10.1145/3351095.3372843
|View full text |Cite
|
Sign up to set email alerts
|

Bias in word embeddings

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
39
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 72 publications
(40 citation statements)
references
References 67 publications
0
39
0
1
Order By: Relevance
“…It has been proven that classifiers trained on biased word embeddings replicate the bias encoded in the word embeddings (Papakyriakopoulos et al, 2020 ). The authors trained a sentiment classifier on word embeddings that had before been proven to be biased.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…It has been proven that classifiers trained on biased word embeddings replicate the bias encoded in the word embeddings (Papakyriakopoulos et al, 2020 ). The authors trained a sentiment classifier on word embeddings that had before been proven to be biased.…”
Section: Discussionmentioning
confidence: 99%
“…Other approaches consider a mitigation at the level of the application that is using the word embeddings. For example, it has been shown that the mitigation at the level of a classifier that was trained on biased word embeddings was efficient (Papakyriakopoulos et al, 2020 ). The authors claim that the classifier learns further associations between the vectors, which are not considered when debiasing at the level of the word embeddings.…”
Section: Discussionmentioning
confidence: 99%
“…On the other hand, Merenda et al (2018) have shown the effectiveness of using messages from potentially abusive-oriented on-line communities to generate so-called hate embeddings. More recently, Papakyriakopoulos et al (2020) have shown that biased word embeddings can be beneficial. We follow the idea of exploiting biased embeddings by creating them using messages from banned communities in Reddit.…”
Section: Hatebert: Re-training Bert With Abusive Online Communitiesmentioning
confidence: 99%
“…Similarly to how long existing stereotypes are deep-rooted in word embeddings (Papakyriakopoulos et al, 2020;Garg et al, 2018), PTLMs have also been shown to recreate stereotypical content due to the nature of their training data (Sheng et al, 2019) Different probing experiments have been proposed to study the drawbacks of PTLMs in areas such as the biomedical domain (Jin et al, 2019), syntax (Hewitt and Manning, 2019;Marvin and Linzen, 2018), semantic and syntactic sentence structures (Tenney et al, 2019), prenomial anaphora (Sorodoc et al, 2020), common-sense (Petroni et al, 2019), gender bias (Kurita et al, 2019), and typicality in judgement (Misra et al, 2021). Except for Hutchinson et al (2020) who examine what words BERT generate in some fill-in-the-blank experiments with regard to people with disabilities, and more recently Nozza et al (2019) who assess hurtful auto-completion by multilingual PTLMs, we are not aware of other strategies designed to estimate toxic content in PTLMs with regard to several social groups.…”
Section: Related Workmentioning
confidence: 99%
“…Bias in social data is a broad concept which involves several issues and formalism (Kiritchenko and Mohammad, 2018;Olteanu et al, 2019;Papakyriakopoulos et al, 2020;Blodgett et al, 2020). For instance, Shah et al (2020) present a framework to predict the origin of different types of bias including label bias (Sap et al, 2019a), selection bias (Garimella et al, 2019;Ousidhoum et al, 2020), model overamplification (Zhao et al, 2017), and semantic bias (Garg et al, 2018).…”
Section: Related Workmentioning
confidence: 99%