2022
DOI: 10.18280/ria.360315
|View full text |Cite
|
Sign up to set email alerts
|

Arabic Language Modeling Based on Supervised Machine Learning

Abstract: Misinformation and misleading actions have appeared as soon as COVID-19 vaccinations campaigns were launched, no matter what the country’s alphabetization level or growing index is. In such a situation, supervised machine learning techniques for classification appears as a suitable solution to model the value & veracity of data, especially in the Arabic language as a language used by millions of people around the world. To achieve this task, we had to collect data manually from SM platforms such as Faceboo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(2 citation statements)
references
References 21 publications
0
1
0
Order By: Relevance
“…In some literature, it is explained that data preprocessing is a must for processing text data in the field of natural language processing. At the preprocessing stage [50], the text will carry out the process stages such as case folding which changes the letters in the data to lowercase, tokenizing aims to separate each word in the data, stopword aims to collect words that are not important and deletes data and stemming aims to make changes to words that affixes become standard words or basic words after that the data is analyzed manually to see the word structure if there are still words that are not good then manual annotations will be made to ensure the data is structured then the data will be divided into 2 columns, namely the text data column and the summary column data with the aim of implementing the T5 model and optimizing parameters with Bayesian optimization.…”
Section: Data Acquisition and Processingmentioning
confidence: 99%
“…In some literature, it is explained that data preprocessing is a must for processing text data in the field of natural language processing. At the preprocessing stage [50], the text will carry out the process stages such as case folding which changes the letters in the data to lowercase, tokenizing aims to separate each word in the data, stopword aims to collect words that are not important and deletes data and stemming aims to make changes to words that affixes become standard words or basic words after that the data is analyzed manually to see the word structure if there are still words that are not good then manual annotations will be made to ensure the data is structured then the data will be divided into 2 columns, namely the text data column and the summary column data with the aim of implementing the T5 model and optimizing parameters with Bayesian optimization.…”
Section: Data Acquisition and Processingmentioning
confidence: 99%
“…In particular, the proposal of a bidirectional encoder representations from transformers (BERT) model has brought revolutionary progress to NLP tasks [12][13][14][15]. Performance has been significantly improved in various NLP tasks using BERT-based pre-training models, such as generative pre-trained transformer (GPT) series of models [16][17][18].…”
Section: Students' Composition Evaluation Model Based On a Natural La...mentioning
confidence: 99%