2020
DOI: 10.1145/3390092
|View full text |Cite
|
Sign up to set email alerts
|

Robust Arabic Text Categorization by Combining Convolutional and Recurrent Neural Networks

Abstract: Text Categorization is an important task in the area of Natural Language Processing (NLP). Its goal is to learn a model that can accurately classify any textual document for a given language into one of a set of predefined categories. In the context of the Arabic language, several approaches have been proposed to tackle this problem, many of which are based on the bag-of-words assumption. Even though these methods usually produce good results for the classification task, they often fail to capture contextual d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(9 citation statements)
references
References 29 publications
0
9
0
Order By: Relevance
“…Recurrent Neural Network (RNN) and its different variants are one of the most preferred models by researchers, because of their capability to model complex underlying relationships in sequential data. Mohamed Seghir Hadj Ameur et al 43 have used bi-directional Gated Recurrent Unit (GRU) to learn dependencies between words for extractive text summarization. Qingyu Zhou et al 44 have applied a 2-layer normalized RNN model on vectors encoded (BERT) to extract summary sentences.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Recurrent Neural Network (RNN) and its different variants are one of the most preferred models by researchers, because of their capability to model complex underlying relationships in sequential data. Mohamed Seghir Hadj Ameur et al 43 have used bi-directional Gated Recurrent Unit (GRU) to learn dependencies between words for extractive text summarization. Qingyu Zhou et al 44 have applied a 2-layer normalized RNN model on vectors encoded (BERT) to extract summary sentences.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Recurrent neural network (RNN) and its different variants are one of the most preferred models by researchers because of their capability to model complex underlying relationships in sequential data. Mohamed Seghir Hadj Ameur et al [72] used a bi-directional gated recurrent unit (GRU) to learn dependencies between words for extractive text summarization. Qingyu Zhou et al [73] applied a twolayer normalized RNN model on vectors encoded (BERT) to extract summary sentences, which are also discussed in [74].…”
Section: Literature Reviewmentioning
confidence: 99%
“…In 2020, Daif et al presented AraDIC [6], the first deep learning framework for Arabic document classification based on image-based characters Ameur et al suggested a hybrid CNN and RNN deep learning model for categorizing Arabic text documents using static, dynamic, and fine-tuned word embedding [3]. The most meaningful representations from the space of Arabic word embedding are automatically learned using a deep learning CNN model.…”
Section: Classificationmentioning
confidence: 99%
“…The Arabic language is the main language of 26 Arab countries (i.e., Arab countries) which possesses many difficulties compared to English. Arabic text analytics are incredibly significant with respect to making our lives easier in many domains such as document text categorization [3], Arabic sentiment analysis [4], and detection of email spam. In fact, the Arabic text faces many challenges as mentioned in [5] such as stemming, dialects, phonology, orthography, and morphology.…”
Section: Introductionmentioning
confidence: 99%