2021
DOI: 10.7717/peerj-cs.559
|View full text |Cite
|
Sign up to set email alerts
|

Investigating cross-lingual training for offensive language detection

Abstract: Platforms that feature user-generated content (social media, online forums, newspaper comment sections etc.) have to detect and filter offensive speech within large, fast-changing datasets. While many automatic methods have been proposed and achieve good accuracies, most of these focus on the English language, and are hard to apply directly to languages in which few labeled datasets exist. Recent work has therefore investigated the use of cross-lingual transfer learning to solve this problem, training a model … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(17 citation statements)
references
References 46 publications
0
17
0
Order By: Relevance
“…Overall, multilingual models trained or finetuned directly on the target languages, i.e. in manyshot settings, are still consistently found to perform best (Aluru et al, 2020;Pelicon et al, 2021). MHC's functional tests are model-agnostic and can be used to evaluate multilingual hate speech detection models trained on any amount of data.…”
Section: Multilingual Hate Speech Detectionmentioning
confidence: 99%
“…Overall, multilingual models trained or finetuned directly on the target languages, i.e. in manyshot settings, are still consistently found to perform best (Aluru et al, 2020;Pelicon et al, 2021). MHC's functional tests are model-agnostic and can be used to evaluate multilingual hate speech detection models trained on any amount of data.…”
Section: Multilingual Hate Speech Detectionmentioning
confidence: 99%
“…Additionally, several works use BERT in zero-shot cross-lingual setups. Pelicon et al (2021) and Nozza (2021) use mBERT, the multilingual version of BERT, without any intermediate steps between source-language training and target-language testing. In this work we use mBERT.…”
Section: Hate Speech Data Scarcity and Cross-lingual Transfermentioning
confidence: 99%
“…A typical form of toxic disinhibition is flaming behavior, which involves using hostile expressions to refer to other users in online communication. Textual features of flaming behavior include harsh language, negative connotations, sexual harassment, and disrespectful expressions ( Pelicon et al, 2021 ). The definition of toxic disinhibition, or toxic behavior, varies based on the users, the communities, and the types of interactions ( Shores et al, 2014 ).…”
Section: Literature Reviewmentioning
confidence: 99%