Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop 2022
DOI: 10.18653/v1/2022.acl-srw.4
|View full text |Cite
|
Sign up to set email alerts
|

Darkness can not drive out darkness: Investigating Bias in Hate SpeechDetection Models

Abstract: This paper is a summary of the work in my PhD thesis. In which, I investigate the impact of bias in NLP models on the task of hate speech detection from three perspectives: explainability, offensive stereotyping bias, and fairness. I discuss the main takeaways from my thesis and how they can benefit the broader NLP community. Finally, I discuss important future research directions. The findings of my thesis suggest that bias in NLP models impacts the task of hate speech detection from all three perspectives. A… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 42 publications
0
1
0
Order By: Relevance
“…Huang et al (2020) argued that whether a statement is considered hate speech depends largely on who the speaker is. Elsafoury (2022) investigated the causal effect of the social and intersectional bias on the performance and unfairness of hate speech detection models. Therefore, some debiasing methods for this task have also been proposed.…”
Section: Monolingual Text Classification and Fairness Researchmentioning
confidence: 99%
“…Huang et al (2020) argued that whether a statement is considered hate speech depends largely on who the speaker is. Elsafoury (2022) investigated the causal effect of the social and intersectional bias on the performance and unfairness of hate speech detection models. Therefore, some debiasing methods for this task have also been proposed.…”
Section: Monolingual Text Classification and Fairness Researchmentioning
confidence: 99%
“…I introduce the systematic offensive stereotyping (SOS) bias and formally define it as "A systematic association in the word embeddings between profanity and marginalized groups of people." (Elsafoury, 2022). I propose a method to measure it and validate it in static (Elsafoury et al, 2022a) and contextual word embeddings (Elsafoury et al, 2022a).…”
Section: The Offensive Stereotyping Bias Perspectivementioning
confidence: 99%
“…I introduce the systematic offensive stereotyping (SOS) bias and formally define it as "A systematic association in the word embeddings between profanity and marginalized groups of people." (Elsafoury, 2022). I propose a method to measure it and validate it in static (Elsafoury et al, 2022a) and contextual word embeddings (Elsafoury et al, 2022a).…”
Section: Contributionsmentioning
confidence: 99%