Proceedings of the 4th Workshop on Gender Bias in Natural Language Processing (GeBNLP) 2022
DOI: 10.18653/v1/2022.gebnlp-1.15
|View full text |Cite
|
Sign up to set email alerts
|

An Empirical Study on the Fairness of Pre-trained Word Embeddings

Abstract: Pre-trained word embedding models are easily distributed and applied, as they alleviate users from the effort to train models themselves. With widely distributed models, it is important to ensure that they do not exhibit undesired behaviour, such as biases against population groups. For this purpose, we carry out an empirical study on evaluating the bias of 15 publicly available, pre-trained word embeddings model based on three training algorithms (GloVe, word2vec, and fastText) with regard to four bias metric… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 46 publications
0
2
0
Order By: Relevance
“…Moreover, as our approach focuses on measuring intrinsic biases of word embedding models (i.e., bias residing in the embedding vectors) [97,98], an interesting avenue of future work are extrinsic (i.e., downstream) tasks to determine the fairness of word embedding models (e.g., co-reference resolution or hate speech detection).…”
Section: Discussionmentioning
confidence: 99%
“…Moreover, as our approach focuses on measuring intrinsic biases of word embedding models (i.e., bias residing in the embedding vectors) [97,98], an interesting avenue of future work are extrinsic (i.e., downstream) tasks to determine the fairness of word embedding models (e.g., co-reference resolution or hate speech detection).…”
Section: Discussionmentioning
confidence: 99%
“…WEAT has traditionally been applied to uncover biases in static word embeddings like FastText and GloVe (Lauscher and Glavaš, 2019;Caliskan et al, 2022;Sesari et al, 2022). However, recent developments in the field have seen a significant shift towards contextualized embeddings, e.g.…”
Section: Bias Sensitivity Evaluationmentioning
confidence: 99%