2021
DOI: 10.1145/3466171
|View full text |Cite
|
Sign up to set email alerts
|

Empirical Evaluation of Shallow and Deep Learning Classifiers for Arabic Sentiment Analysis

Abstract: This work presents a detailed comparison of the performance of deep learning models such as convolutional neural networks, long short-term memory, gated recurrent units, their hybrids, and a selection of shallow learning classifiers for sentiment analysis of Arabic reviews. Additionally, the comparison includes state-of-the-art models such as the transformer architecture and the araBERT pre-trained model. The datasets used in this study are multi-dialect Arabic hotel and book review datasets, which are some of… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 14 publications
(16 citation statements)
references
References 56 publications
0
15
1
Order By: Relevance
“…Table 14 compares the best-achieved results of balanced and unbalanced HARD datasets using the proposed models after emoji and emoticon replacement and using the features with the results of other studies in the literature that predicted user opinions based on five levels of ratings. For the balanced HARD data, using the EDLB-MLP model with the total weight of emojis and emoticons, the proposed classifier has achieved an increase of 3.21% in accuracy over Nassif et al's [20] top five models: random forest, convolutional neural network, decision tree, gated recurrent unit, and bi-directional recurrent neural network. The percentage increase formula is derived from the concept of percentage increase, as follows: [(New Accuracy − Old Accuracy)/Old Accuracy] × 100.…”
Section: Resultsmentioning
confidence: 93%
See 1 more Smart Citation
“…Table 14 compares the best-achieved results of balanced and unbalanced HARD datasets using the proposed models after emoji and emoticon replacement and using the features with the results of other studies in the literature that predicted user opinions based on five levels of ratings. For the balanced HARD data, using the EDLB-MLP model with the total weight of emojis and emoticons, the proposed classifier has achieved an increase of 3.21% in accuracy over Nassif et al's [20] top five models: random forest, convolutional neural network, decision tree, gated recurrent unit, and bi-directional recurrent neural network. The percentage increase formula is derived from the concept of percentage increase, as follows: [(New Accuracy − Old Accuracy)/Old Accuracy] × 100.…”
Section: Resultsmentioning
confidence: 93%
“…Thus, the balanced HARD dataset has four rating levels, while the unbalanced version has five. In [20], the authors developed different deep and machine learning models for opinion mining in the reviews' domain. They used a subset of the HARD dataset, comprising 62,500 random reviews, to predict the five rating levels.…”
Section: Arabic Dataset and Opinion Mining Methodsmentioning
confidence: 99%
“…Natural Language Processing (NLP) for Arabic language has become a very interesting and challenging topic for researchers with its various topics and tasks [ 10 ]. In addition to the fake news detection and spam detection, there are many important and related tasks to begin with such as Arabic sentiment analysis (ASA), question answering system in Arabic language, etc.…”
Section: Related Workmentioning
confidence: 99%
“…For RNN, we experimented with both the Bidirectional Long Short-Term Memory (BiLSTM) layer and Bidirectional Gated Recurrent Unit (BiGRU). We selected these three models because of their good performance on similar tasks (Al Qadi et al 2019;Al Qadi et al 2020;Elnagar et al 2020;Nassif, Darya, and Elnagar 2021a;El Rifai, Al Qadi, and Elnagar 2022).…”
Section: Extrinsic Assessmentmentioning
confidence: 99%