Findings of the Association for Computational Linguistics: ACL 2022 2022
DOI: 10.18653/v1/2022.findings-acl.32
|View full text |Cite
|
Sign up to set email alerts
|

Dynamically Refined Regularization for Improving Cross-corpora Hate Speech Detection

Abstract: Warning: this paper contains content that may be offensive and distressing.Hate speech classifiers exhibit substantial performance degradation when evaluated on datasets different from the source. This is due to learning spurious correlations between words that are not necessarily relevant to hateful language, and hate speech labels from the training corpus. Previous work has attempted to mitigate this problem by regularizing specific terms from pre-defined static dictionaries. While this has been demonstrated… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 18 publications
0
2
0
Order By: Relevance
“…Current HS models often suffer from spurious correlations (Ramponi and Tonelli, 2022;Bose et al, 2022;Wang et al, 2022): they tend to utilize prediction patterns that hold for the majority examples but do not hold in general. This might cause the model to be biased in applications (Ramponi and Tonelli, 2022;Gardner et al, 2021).…”
Section: Identifying Spurious Correlationsmentioning
confidence: 99%
“…Current HS models often suffer from spurious correlations (Ramponi and Tonelli, 2022;Bose et al, 2022;Wang et al, 2022): they tend to utilize prediction patterns that hold for the majority examples but do not hold in general. This might cause the model to be biased in applications (Ramponi and Tonelli, 2022;Gardner et al, 2021).…”
Section: Identifying Spurious Correlationsmentioning
confidence: 99%
“…Chrysostomou and Aletras (2022a) further show that post-hoc explanation methods might not provide faithful explanations in out-of-domain settings. The contemporaneous work by Attanasio et al (2022) and Bose et al (2022) reduce lexical overfitting automatically with entropy-based attentions and feature attributions, respectively. While cross-domain classification performance across different datasets is not studied in the former, the latter needs some labeled target instances to identify the over-fitted terms.…”
Section: Introductionmentioning
confidence: 99%