2022
DOI: 10.1145/3555088
|View full text |Cite
|
Sign up to set email alerts
|

Is Your Toxicity My Toxicity? Exploring the Impact of Rater Identity on Toxicity Annotation

Abstract: Machine learning models are commonly used to detect toxicity in online conversations. These models are trained on datasets annotated by human raters. We explore how raters' self-described identities impact how they annotate toxicity in online comments. We first define the concept of Specialized Rater Pools: rater pools formed based on raters' self-described identities, rather than at random. We formed three such rater pools for this study - specialized rater pools of raters from the U.S. who identify as Africa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
29
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 32 publications
(30 citation statements)
references
References 17 publications
1
29
0
Order By: Relevance
“…In this domain, where labelers are annotating portraits of humans, we can conclude that labeler bias exists, depends on labeler demographics, and can be explained using stereotype content [20]. These results are in line with recent findings in CSCW by Goyal et al [24] demonstrating that toxicity labels for online content are influenced by labeler self-identification.…”
Section: Labelers Exhibit Biassupporting
confidence: 86%
See 2 more Smart Citations
“…In this domain, where labelers are annotating portraits of humans, we can conclude that labeler bias exists, depends on labeler demographics, and can be explained using stereotype content [20]. These results are in line with recent findings in CSCW by Goyal et al [24] demonstrating that toxicity labels for online content are influenced by labeler self-identification.…”
Section: Labelers Exhibit Biassupporting
confidence: 86%
“…Prior research has found that even highly experienced labelers fail to produce unbiased labels [29]. Perhaps most relevant to our study is recent work in CSCW by Goyal et al [24] on rater identity. They found that rater identity (i.e., African American, LGBTQ, or neither) significantly influenced how raters annotated toxicity in online comments.…”
Section: Bias In Machine Learningmentioning
confidence: 83%
See 1 more Smart Citation
“…For example, lack of familiarity with the dialect of the content's author can lead to biased decisions [41,10]. Whether the moderator has themself been a target of hate speech [29], or whether their own demographic identity aligns with that being targeted in content they are reviewing [19] can also impact their decisions. Thus, once differences in judging behavior among moderators is acknowledged and accepted, it creates a space for matching different groups of moderators to different content types, based on moderator background (which can be collected via an on-boarding questionnaire).…”
Section: Injecting Prior Informationmentioning
confidence: 99%
“…We demonstrate the efficacy of our methods via both theoretical analysis ( §2.3) and empirical improvement on synthetic and real-world datasets ( §3 and §4). The latter extends beyond the classic assumption of universal, objective truth to consider recent advocacy for recognizing subjective, community-based gold standards [45,29,17,19].…”
Section: Introductionmentioning
confidence: 99%