2019
DOI: 10.48550/arxiv.1903.10561
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On Measuring Social Biases in Sentence Encoders

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
56
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 35 publications
(57 citation statements)
references
References 0 publications
1
56
0
Order By: Relevance
“…Safety and safety of dialog models: Inappropriate and unsafe risks and behaviors of language models have been extensively discussed and studied in previous works (e.g., [53,54]). Issues encountered include toxicity (e.g., [55,56,57]), bias (e.g., [58,59,60,61,62,63,64,65,66,67,68,69,70,71,72]), and inappropriately revealing personally identifying information (PII) from training data [73]. Weidinger et al [54] identify 21 risks associated with large-scale language models and discuss the points of origin for these risks.…”
Section: Related Workmentioning
confidence: 99%
“…Safety and safety of dialog models: Inappropriate and unsafe risks and behaviors of language models have been extensively discussed and studied in previous works (e.g., [53,54]). Issues encountered include toxicity (e.g., [55,56,57]), bias (e.g., [58,59,60,61,62,63,64,65,66,67,68,69,70,71,72]), and inappropriately revealing personally identifying information (PII) from training data [73]. Weidinger et al [54] identify 21 risks associated with large-scale language models and discuss the points of origin for these risks.…”
Section: Related Workmentioning
confidence: 99%
“…Here, we investigate how the sensitivity varies across the range of toxicity scores, which gives important clues about the desirability of biases. Instead of using naturally occurring sentences (like (Prabhakaran et al, 2019)) which would intro-duce numerous unknown correlational effects, we followed Hutchinson et al (2020) and May et al (2019) by building a set of 33 template sentences (see Appendix 1) that correspond to a range of toxicity scores. We list a few examples of the templates below:…”
Section: Methodsmentioning
confidence: 99%
“…Racial Bias: In language generation using OpenAI's GTP-2 model, Sheng et al [42] show that there are more negative associations of the black population when conditioning on context related to respect and occupation. Another study adapts the Sentence Encoder Association Test (SEAT) [34] to analyze potential biases encoded in BERT and GPT-2 with respect to gender, race, and the intersectional identities (gender + race). The empirical analysis shows that BERT has the highest proportion of bias on the race and intersectional tests performed among all contextual word models [46].…”
Section: Fairness/bias In Language Modelsmentioning
confidence: 99%