2022
DOI: 10.11591/ijai.v11.i1.pp379-387
|View full text |Cite
|
Sign up to set email alerts
|

AraBERT transformer model for Arabic comments and reviews analysis

Abstract: Arabic language is rich and complex in terms of word morphology compared to other Latin languages. Recently, natural language processing (NLP) field emerges with many researches targeting Arabic language understanding (ALU). In this context, this work presents our developed approach based on the Arabic bidirectional encoder representations from transformers (AraBERT) model where the main required steps are presented in detail. We started by the input text pre-processing, which is, then, segmented using the Far… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 18 publications
0
3
0
Order By: Relevance
“…12 transformer-based encoders are there in a BERT base model. The output of this model will be a vector of size 768, which can be given to a classification layer for the task of classification [25]- [31].…”
Section: Bertmentioning
confidence: 99%
“…12 transformer-based encoders are there in a BERT base model. The output of this model will be a vector of size 768, which can be given to a classification layer for the task of classification [25]- [31].…”
Section: Bertmentioning
confidence: 99%
“…Arabic is a widely spoken and written language with a significant presence in the online world. Researchers in the Arabic world have started to focus on creating resources and language models for the Arabic language; examples of Arabic language models include AraBERT [11], ARBERT, MARBET [12], and CAMeLBERT [13], all of which focus on Modern Standard Arabic (MSA). In addition, there are models that cover Arabic dialects for specific countries.…”
Section: Natural Language Processing (Nlp) Of Kuwaiti Dialectmentioning
confidence: 99%
“…Several recent studies [12], [13] have trained the bidirectional encoder representations from transformers (BERT) model on Wikipedia and Oscar datasets for the Arabic language. In addition to that, several recent studies [14], [15] have fine-tuned the Arabic BERT model [13] for downstream task SA. the drawback identified from the analysis of existing literature are: i) models not tested on different datasets; ii) some models ignore the context meaning of the sentence; iii) the model using context like BERT fined-tuned using general pre-trained models that affect models performance; and iv) there is room for improvement for reported prediction accuracy.…”
Section: Introductionmentioning
confidence: 99%