2024
DOI: 10.1145/3643829
|View full text |Cite
|
Sign up to set email alerts
|

“HOT” ChatGPT: The Promise of ChatGPT in Detecting and Discriminating Hateful, Offensive, and Toxic Comments on Social Media

Lingyao Li,
Lizhou Fan,
Shubham Atreja
et al.

Abstract: Harmful textual content is pervasive on social media, poisoning online communities and negatively impacting participation. A common approach to this issue is developing detection models that rely on human annotations. However, the tasks required to build such models expose annotators to harmful and offensive content and may require significant time and cost to complete. Generative AI models have the potential to understand and detect harmful textual content. We used ChatGPT to investigate this potential and co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2024
2024
2025
2025

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 26 publications
(2 citation statements)
references
References 75 publications
0
2
0
Order By: Relevance
“…Researchers are examining the potential of GPT-3 models for the classification of hateful content (Chiu, Collins and Alexander 2022;Wang and Chang 2022;Huang, Kwak and An 2023). Li et al (2023) conduct extensive prompting experiments and compare the performance of ChatGPT to that of crowdworkers for the task of classifying texts as hateful, offensive or toxic (HOT). They find that ChatGPT achieves an accuracy of roughly 80% when compared to crowdworkers' annotations.…”
Section: Text Classification With Prompt-based Generative Modelsmentioning
confidence: 99%
“…Researchers are examining the potential of GPT-3 models for the classification of hateful content (Chiu, Collins and Alexander 2022;Wang and Chang 2022;Huang, Kwak and An 2023). Li et al (2023) conduct extensive prompting experiments and compare the performance of ChatGPT to that of crowdworkers for the task of classifying texts as hateful, offensive or toxic (HOT). They find that ChatGPT achieves an accuracy of roughly 80% when compared to crowdworkers' annotations.…”
Section: Text Classification With Prompt-based Generative Modelsmentioning
confidence: 99%
“…Researchers are examining the potential of GPT-3 models for the classification of hateful content (Chiu, Collins and Alexander 2022;Wang and Chang 2022;Huang, Kwak and An 2023). Li et al (2023) conduct extensive prompting experiments and compare the performance of ChatGPT to that of crowdworkers for the task of classifying texts as hateful, offensive or toxic (HOT). They find that ChatGPT achieves an accuracy of roughly 80% when compared to crowdworkers' annotations.…”
Section: Text Classification With Prompt-based Generative Modelsmentioning
confidence: 99%