Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2023
DOI: 10.18653/v1/2023.acl-long.227
|View full text |Cite
|
Sign up to set email alerts
|

Social-Group-Agnostic Bias Mitigation via the Stereotype Content Model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 0 publications
0
2
0
Order By: Relevance
“…Though modern approaches have made great advances in democratizing LLM training, most builders don't have a need to pretrain their own LLMs, opting to, at most, fine-tune them. Rather than hope that an LLM is unbiased after pretraining, many researchers have discussed the utility in having a separate general debiasing step to account for any unintended associations stemming from pretraining Omrani et al, 2023;. Relatively less explored is the complementary requirement of augmenting LLMs with the awareness and ability to abide by sociocultural norms.…”
Section: Human-centered Nlpmentioning
confidence: 99%
“…Though modern approaches have made great advances in democratizing LLM training, most builders don't have a need to pretrain their own LLMs, opting to, at most, fine-tune them. Rather than hope that an LLM is unbiased after pretraining, many researchers have discussed the utility in having a separate general debiasing step to account for any unintended associations stemming from pretraining Omrani et al, 2023;. Relatively less explored is the complementary requirement of augmenting LLMs with the awareness and ability to abide by sociocultural norms.…”
Section: Human-centered Nlpmentioning
confidence: 99%
“…They disentangle po-tentially correlated concepts by projecting representations into orthogonal subspaces, thus removing discriminatory correlation bias (Dev et al 2021;Kaneko and Bollegala 2021). Group-specific subspace projection requires prior group knowledge, some work (Ungless et al 2022;Omrani et al 2023) projects representations to stereotype content models (SCM) (Fiske et al 2002) that rely on theoretical understanding of social stereotypes to define bias subspaces, thus breaking the limitations of prior knowledge.…”
Section: In-processingmentioning
confidence: 99%