2022
DOI: 10.1007/978-3-030-96957-8_33
|View full text |Cite
|
Sign up to set email alerts
|

Testing the Generalization of Neural Language Models for COVID-19 Misinformation Detection

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(4 citation statements)
references
References 22 publications
0
4
0
Order By: Relevance
“…The study from 2021 [53] evaluated fifteen transformer-based models to detect misinformation in several COVID-19 datasets, which are COVID-CQ [54], CoAID [55], Re-COVery [56], CMU-MisCov19 [57], and COVID19FN [58]. In this study, tokenizers and models tailored to COVID-19 data did not outperform general-purpose tokenizers and models, according to the researchers.…”
Section: Discussionmentioning
confidence: 94%
“…The study from 2021 [53] evaluated fifteen transformer-based models to detect misinformation in several COVID-19 datasets, which are COVID-CQ [54], CoAID [55], Re-COVery [56], CMU-MisCov19 [57], and COVID19FN [58]. In this study, tokenizers and models tailored to COVID-19 data did not outperform general-purpose tokenizers and models, according to the researchers.…”
Section: Discussionmentioning
confidence: 94%
“…Another study using the ReCOVery data set for model development explored the use of multiple languages for fake news detection to improve model performance [ 67 ]. Finally, Wahle et al [ 68 ] used the ReCOVery data set as 1 of 6 COVID-19 misinformation data sets to evaluate the performance of 15 transformer-based ML models to determine the generalizability of different transformer models. Differing from the aforementioned studies, we were able to demonstrate that the use of readability, text characteristics, sentiment, and lexical categories can improve upon the original ReCOVery data set baseline models [ 17 ].…”
Section: Discussionmentioning
confidence: 99%
“…BART is a two-stage denoising autoencoder that corrupts its input text and reconstructs it through a sequence-to-sequence model. We chose BART because of its ability to perform a wide range of downstream tasks, such as paraphrase detection (Wahle et al, 2022b), fake news identification (Wahle et al, 2022a), and text summarization (Lewis et al, 2020). Additionally, in our preliminary experiments, BART also performed better than other candidate models such as PEGASUS (Zhang et al, 2020) and T5 (Raffel et al, 2020) (comparison in Tables 9 and 10 in appendix B).…”
Section: Methodsmentioning
confidence: 99%