Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.551
|View full text |Cite
|
Sign up to set email alerts
|

ARBERT & MARBERT: Deep Bidirectional Transformers for Arabic

Abstract: Pre-trained language models (LMs) are currently integral to many natural language processing systems. Although multilingual LMs were also introduced to serve many languages, these have limitations such as being costly at inference time and the size and diversity of non-English data involved in their pre-training. We remedy these issues for a collection of diverse Arabic varieties by introducing two powerful deep bidirectional transformer-based models, ARBERT and MARBERT. To evaluate our models, we also introdu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
122
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 193 publications
(122 citation statements)
references
References 36 publications
0
122
0
Order By: Relevance
“…Only two studies, however, attempted to use PLMs and TL for the task of Arabic emotion classification. The first model was a BERT-based model named MARBERT (Abdul-Mageed et al, 2020 ), which was trained on tweets, and social meaning tasks were considered in the fine-tuning. The social meaning tasks included emotion detection.…”
Section: Emotion Analysis Of Arabic Tweetsmentioning
confidence: 99%
See 1 more Smart Citation
“…Only two studies, however, attempted to use PLMs and TL for the task of Arabic emotion classification. The first model was a BERT-based model named MARBERT (Abdul-Mageed et al, 2020 ), which was trained on tweets, and social meaning tasks were considered in the fine-tuning. The social meaning tasks included emotion detection.…”
Section: Emotion Analysis Of Arabic Tweetsmentioning
confidence: 99%
“…Textual data on Twitter are frequently dialectal, with dialects lacking spelling norms and informal in nature, and may include emojis, hashtags, and user mentions. Only QARiB (Abdelali et al, 2021 ) and MARBERT (Abdul-Mageed et al, 2020 ), as mentioned above, used tweets to train transformer models and fine-tune for emotion classification tasks, and both of them are BERT-based. Moreover, even when testing on informal text, increasing the variety of training data by including both formal (MSA) and informal (AD) text is better than using informal text alone (Abdelali et al, 2021 ).…”
Section: Emotion Analysis Of Arabic Tweetsmentioning
confidence: 99%
“…Transformers have recently become popular in NLP and text categorization. The transformer model is a deep neural network architecture based completely on the attention mechanism, replacing the recurrent layers with auto encoder-decoder architectures with special called multi-head self-attention layers [41].…”
Section: Machine Learning Algorithmsmentioning
confidence: 99%
“…El Mahdaouy et al [64] approached the same shared task using an end-to-end multitask learning model based on the MarBERT [65] language model. Duwairi et al [66] investigated the ability of CNN, CNN-LSTM, and BiLSTM-CNN networks to detect hateful content on social media.…”
Section: Hate and Offensive Speech Detection In Arabicmentioning
confidence: 99%
“…Each word in the inputted sequence will be mapped to an embedding vector generated from summing up its corresponding word, segment, and positional embeddings. To learn shared global contextual representations across all corpora, the shared part is employed to fine-tune the weights of a pre-trained multilayer bidirectional transformer encoder such as AraBERT [41], and MarBERT [65]. The AraBERT and MarBERT are transformer encoders consisting of a self-attention mechanism to learn the contextual representations for each inputted word.…”
Section: Proposed Modelmentioning
confidence: 99%