Proceedings of the Fourth Workshop on Online Abuse and Harms 2020
DOI: 10.18653/v1/2020.alw-1.3
|View full text |Cite
|
Sign up to set email alerts
|

Using Transfer-based Language Models to Detect Hateful and Offensive Language Online

Abstract: Distinguishing hate speech from non-hate offensive language is challenging, as hate speech not always includes offensive slurs and offensive language not always express hate. Here, four deep learners based on the Bidirectional Encoder Representations from Transformers (BERT), with either general or domain-specific language models, were tested against two datasets containing tweets labelled as either 'Hateful', 'Normal' or 'Offensive'. The results indicate that the attention-based models profoundly confuse hate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(11 citation statements)
references
References 37 publications
1
10
0
Order By: Relevance
“…While BERT has been pre-trained using vast amounts of textual data, namely, all of Wikipedia and Google Books, the word encodings produced by BERT can benefit from further adaptation to the target domain (Gururangan et al, 2020). In a recent work, Isaksen and Gambäck (2020) applied BERT to the task of hate detection. They report results of finetuning and evaluating the base and large variants of BERT (BERT-base and BERT-large) on each of the DV and FN datasets.…”
Section: Hate Speech Detection With Pre-trained Language Modelsmentioning
confidence: 99%
“…While BERT has been pre-trained using vast amounts of textual data, namely, all of Wikipedia and Google Books, the word encodings produced by BERT can benefit from further adaptation to the target domain (Gururangan et al, 2020). In a recent work, Isaksen and Gambäck (2020) applied BERT to the task of hate detection. They report results of finetuning and evaluating the base and large variants of BERT (BERT-base and BERT-large) on each of the DV and FN datasets.…”
Section: Hate Speech Detection With Pre-trained Language Modelsmentioning
confidence: 99%
“…Besides, SOTA approaches like deep learning (Badjatiya et al, 2017) and transformers models (Isaksen and Gambäck, 2020) are applied in the hate speech detection and toxic posts classification. However, these models only classify based on the whole posts or documents.…”
Section: Related Workmentioning
confidence: 99%
“…However, the authors did not mention the annotation process and the method for evaluating the quality of the dataset. Besides, on the hate speech detection problem, many state-of-the-art models give optimistic results such as deep learning models [4] and transformer language models [20]. Those models require largescale annotated datasets, which is a challenge for low-resource languages like Vietnamese.…”
Section: Related Workmentioning
confidence: 99%