2022
DOI: 10.3390/healthcare10081478
|View full text |Cite
|
Sign up to set email alerts
|

Comparison of Pretraining Models and Strategies for Health-Related Social Media Text Classification

Abstract: Pretrained contextual language models proposed in the recent past have been reported to achieve state-of-the-art performances in many natural language processing (NLP) tasks, including those involving health-related social media data. We sought to evaluate the effectiveness of different pretrained transformer-based models for social media-based health-related text classification tasks. An additional objective was to explore and propose effective pretraining strategies to improve machine learning performance on… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 52 publications
1
9
0
Order By: Relevance
“…A number of included papers used more than one dataset to evaluate their classifiers in parallel, showing differences between the evaluations scores of the same algorithm or architecture and therefore making it hard to estimate how each algorithm would perform in real-life, with potentially new or evolving data. [57][58][59] In total, 15 papers described qualitative or practical evaluations of their algorithms. In the most cases, the qualitative evaluation completely replaced quantitative analysis.…”
Section: Evaluation Of Algorithmsmentioning
confidence: 99%
“…A number of included papers used more than one dataset to evaluate their classifiers in parallel, showing differences between the evaluations scores of the same algorithm or architecture and therefore making it hard to estimate how each algorithm would perform in real-life, with potentially new or evolving data. [57][58][59] In total, 15 papers described qualitative or practical evaluations of their algorithms. In the most cases, the qualitative evaluation completely replaced quantitative analysis.…”
Section: Evaluation Of Algorithmsmentioning
confidence: 99%
“…As an example, Qasim et al 15 detected the fake news of COVID-19 by tuning BERT, RoBERTa, and XLNet. Guo et al 16 evaluated six types of PLMs on social media health-related text classification tasks. An elaborate experiment conducted by Lu et al 40 verified that BERT performed best in all scenarios for classifying the presence or the absence of 16 diseases from patient discharge summaries.…”
Section: Related Workmentioning
confidence: 99%
“… 13 , 14 In the early stages of exploiting such knowledge for the MTC, researchers added additional classifiers on the top of PLMs and fine-tuned both of them using task-specific objective functions. 15 , 16 However, one major problem in such a “fine-tuning” paradigm is that additional parameters are introduced when tuning the extra classifiers. Therefore, how to bridge the gap between pre-training objectives and classification tasks has been considered as a critical factor in making a PLM more suitable for MTC.…”
Section: Introductionmentioning
confidence: 99%
“…In other words, a weight tells how much the input affects the output. Biases, which are always the same, are an extra input for the next layer that will always be 1 [33] [34].…”
Section: Weighted Average or Weighted Sum Ensemblementioning
confidence: 99%
“…www.ijacsa.thesai.org transferring knowledge from distinct but related source domains. ResNet50V2, ResNet101V2', 'MobileNetV3Large', 'MobileNetV3Small', 'MobileNet', 'EfficientNetB0', 'EfficientNetB1', and 'EfficientNetB2' are examples of pretrained networks [33]…”
mentioning
confidence: 99%