2019
DOI: 10.1007/978-3-030-36687-2_77
|View full text |Cite
|
Sign up to set email alerts
|

A BERT-Based Transfer Learning Approach for Hate Speech Detection in Online Social Media

Abstract: Generated hateful and toxic content by a portion of users in social media is a rising phenomenon that motivated researchers to dedicate substantial efforts to the challenging direction of hateful content identification. We not only need an efficient automatic hate speech detection model based on advanced machine learning and natural language processing, but also a sufficiently large amount of annotated data to train a model. The lack of a sufficient amount of labelled hate speech data, along with the existing … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

6
208
0
2

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 264 publications
(216 citation statements)
references
References 21 publications
6
208
0
2
Order By: Relevance
“…To define automated methods with a promising performance for hate speech detection in social media, Natural Language Processing (NLP) has been used jointly with classic Machine Learning (ML) [2][3][4] and Deep Learning (DL) techniques [6,15,16]. The majority of contributions in classic supervised machine learning-based methods, for hate speech detection, rely on different text mining-based features or user-based and platform-based metadata [4,17,18], which require them to define an applicable feature extraction method and prevent them to generalize their approach to new datasets and platforms.…”
Section: Disclaimermentioning
confidence: 99%
See 3 more Smart Citations
“…To define automated methods with a promising performance for hate speech detection in social media, Natural Language Processing (NLP) has been used jointly with classic Machine Learning (ML) [2][3][4] and Deep Learning (DL) techniques [6,15,16]. The majority of contributions in classic supervised machine learning-based methods, for hate speech detection, rely on different text mining-based features or user-based and platform-based metadata [4,17,18], which require them to define an applicable feature extraction method and prevent them to generalize their approach to new datasets and platforms.…”
Section: Disclaimermentioning
confidence: 99%
“…Although some deep neural network models such as Convolutional Neural Networks (CNNs) [16], Long Short-Term Memory Networks (LSTMs) [6], etc., have been employed to enhance the performance of hate speech detection tools, the requirement of a sufficient amount of labeled data and the inability of methods to be generalized have remained as open challenges. To address these limitations some transfer learning methods are proposed recently [15,19]. However these methods enhanced the performance of hate speech detection models, they did not address the existing bias in data and algorithm.…”
Section: Disclaimermentioning
confidence: 99%
See 2 more Smart Citations
“…The deep learning methods can be roughly divided into two categories: one focuses on front-end processing which optimizes the word embedding technology, and the other on mid-end processing which usually uses simple word or character based embedding technology and pays more attention to the middle neural networks processing. The most famous methods focused on front-end processing are Embeddings from Language Models (ELMo) [6] [13], which trains word vectors with context, and Bidirectional Encoder Representation from Transformers (BERT) [14] [15]. BERT is the first deeply bidirectional, unsupervised language representation from unlabeled text by jointly conditioning on both left and right context in all layers.…”
Section: B State Of the Art In Deep Learningmentioning
confidence: 99%