2017
DOI: 10.1007/978-3-319-57351-9_6
|View full text |Cite
|
Sign up to set email alerts
|

The Impact of Toxic Language on the Health of Reddit Communities

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
33
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 52 publications
(33 citation statements)
references
References 4 publications
0
33
0
Order By: Relevance
“…• The Reddit 13 dataset is composed of over 880,000 comments taken from a wide range of subreddits and annotated a few years ago by the Community Sift tool developed by Two Hat Security 14 . This toxicity detection tool, which was used in previous research on toxicity as well (Mohan et al, 2017), uses over 1 million n-gram rules in order to normalize then categorize each message into one of eight risk levels for a wide array of different categories, 0 to 3 being super-safe to questionable, 4 being unknown and 5 to 7 being mild to severe. In our case, we consider the scores assigned to each message in five categories, namely bullying, fighting, sexting, vulgarity, and racism.…”
Section: Toxicity Detectionmentioning
confidence: 99%
“…• The Reddit 13 dataset is composed of over 880,000 comments taken from a wide range of subreddits and annotated a few years ago by the Community Sift tool developed by Two Hat Security 14 . This toxicity detection tool, which was used in previous research on toxicity as well (Mohan et al, 2017), uses over 1 million n-gram rules in order to normalize then categorize each message into one of eight risk levels for a wide array of different categories, 0 to 3 being super-safe to questionable, 4 being unknown and 5 to 7 being mild to severe. In our case, we consider the scores assigned to each message in five categories, namely bullying, fighting, sexting, vulgarity, and racism.…”
Section: Toxicity Detectionmentioning
confidence: 99%
“…For this research, we chose the popular GloVe vectors from the SpaCy, a free, open-source library for NLP in Python. 17 Other research papers deploying GloVe for online hate detection include, for example, Mishra et al [15] and Kshirsagar et al [90] The GloVe model we apply is publicly available 18 and contains 685k keys and 20k unique vectors with 300 dimensions, trained on Common Crawl datasets. 19…”
Section: Word Embeddingsmentioning
confidence: 99%
“…Even though hate has been observed as a problem in multiple online social media platforms, including Reddit, YouTube, Wikipedia, Twitter, and so on [5,7,[16][17][18], apart from a few exploratory studies [15,19], there is a lack of development and testing of models using data from multiple social media platforms. Instead, studies tend to focus on one platform.…”
mentioning
confidence: 99%
“…Toxic commenting has also been found prevalent in general online discussion forums, news websites, and social media platforms. The existing research deals with multiple aspects, such as detection and classification of toxicity [9][10][11], assessing its impact on online communities [12,13], types of toxicity such as cyberbullying and trolling [2,14], and means of defusing online toxicity [15]. To approach toxicity, researchers have investigated multiple social media platforms, such as Twitter, YouTube, Facebook, and Reddit [7,11], as well as comments in online discussion forums and news websites [16].…”
Section: Introductionmentioning
confidence: 99%
“…In the present time, news channels cannot isolate themselves from the audience reactions, but analyzing these reactions is important to understand the various sources of digital bias and to form an analytical relationship to the audience. Finally, the betterment of online experiences by mitigating online toxicity is a matter of societal impact, as toxic conversations impact nearly all online users across social media platforms [10,12,25].…”
Section: Introductionmentioning
confidence: 99%