2021
DOI: 10.1038/s41598-021-01487-w
|View full text |Cite
|
Sign up to set email alerts
|

Dynamics of online hate and misinformation

Abstract: Online debates are often characterised by extreme polarisation and heated discussions among users. The presence of hate speech online is becoming increasingly problematic, making necessary the development of appropriate countermeasures. In this work, we perform hate speech detection on a corpus of more than one million comments on YouTube videos through a machine learning model, trained and fine-tuned on a large set of hand-annotated data. Our analysis shows that there is no evidence of the presence of “pure h… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

7
34
0
1

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
1

Relationship

4
5

Authors

Journals

citations
Cited by 61 publications
(42 citation statements)
references
References 43 publications
7
34
0
1
Order By: Relevance
“…However, note that the level of agreement between the best model and the annotators is very close to the inter-annotator agreement. This result is comparable to other related datasets, where the annotation task is subjective and it is unrealistic to expect perfect agreement between the annotators [ 36 , 37 ]. If one accepts an inherent ambiguity of the hate speech classification task, there is very little room for improvement of the binary classification model.…”
Section: Resultssupporting
confidence: 83%
“…However, note that the level of agreement between the best model and the annotators is very close to the inter-annotator agreement. This result is comparable to other related datasets, where the annotation task is subjective and it is unrealistic to expect perfect agreement between the annotators [ 36 , 37 ]. If one accepts an inherent ambiguity of the hate speech classification task, there is very little room for improvement of the binary classification model.…”
Section: Resultssupporting
confidence: 83%
“… 2016 ; Cinelli et al. 2021 ). However, the performance can be improved if user-related context is taken into account (Gao and Huang 2017 ; Fehn Unsvåg and Gambäck 2018 ).…”
Section: Discussionmentioning
confidence: 99%
“…We used the Italian and Slovenian models in two previous analytical studies on hate speech in social media. The Italian model was used in a work investigating relationships between hate speech and misinformation sources on the Italian YouTube [4]. The Slovenian model was used to perform an analysis on the evolution of retweet communities, hate speech and topics on the Slovenian Twitter during 2018-2020 [8][9][10].…”
Section: Model Training and Evaluationmentioning
confidence: 99%