Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.293
|View full text |Cite
|
Sign up to set email alerts
|

Dynamically Disentangling Social Bias from Task-Oriented Representations with Adversarial Attack

Abstract: Representation learning is widely used in NLP for a vast range of tasks. However, representations derived from text corpora often reflect social biases. This phenomenon is pervasive and consistent across different neural models, causing serious concern. Previous methods mostly rely on a pre-specified, user-provided direction or suffer from unstable training. In this paper, we propose an adversarial disentangled debiasing model to dynamically decouple social bias attributes from the intermediate representations… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
2
1

Relationship

2
8

Authors

Journals

citations
Cited by 18 publications
(7 citation statements)
references
References 29 publications
0
7
0
Order By: Relevance
“…As there is still a huge gap in model performance between unseen and seen slots, in future work, we will focus on improving the performance on unseen slots while maintaining the performance on seen slots. Representation disentanglement (Wang et al, 2021) can be used to disentangle domain-specific and domain-shared knowledge in the source domain, then we can preserve domainshared knowledge and focus on establishing the relation of domain-specific knowledge between the source and the target domain.…”
Section: Conclusion and Discussionmentioning
confidence: 99%
“…As there is still a huge gap in model performance between unseen and seen slots, in future work, we will focus on improving the performance on unseen slots while maintaining the performance on seen slots. Representation disentanglement (Wang et al, 2021) can be used to disentangle domain-specific and domain-shared knowledge in the source domain, then we can preserve domainshared knowledge and focus on establishing the relation of domain-specific knowledge between the source and the target domain.…”
Section: Conclusion and Discussionmentioning
confidence: 99%
“…Several works further investigate the use of adversarial training for removing demographic information from neural/PLM-based text classifiers (Elazar and Goldberg, 2018;Barrett et al, 2019;Wang et al, 2021). Notably, Han et al (2021b) show the benefit of having an ensemble of orthogonal adversaries.…”
Section: Fairness and Bias Mitigation In Nlpmentioning
confidence: 99%
“…In order to enhance the robustness of the model, we introduce adversarial training [15] [16] to reduce the influence of domain noise. First, x i is encoded to obtain the representation h i , and the negative log-likelihood (NLL) loss is calculated as follows:…”
Section: Adversarial Learningmentioning
confidence: 99%