2021
DOI: 10.1155/2021/6633213
|View full text |Cite
|
Sign up to set email alerts
|

ABioNER: A BERT‐Based Model for Arabic Biomedical Named‐Entity Recognition

Abstract: The web is being loaded daily with a huge volume of data, mainly unstructured textual data, which increases the need for information extraction and NLP systems significantly. Named-entity recognition task is a key step towards efficiently understanding text data and saving time and effort. Being a widely used language globally, English is taking over most of the research conducted in this field, especially in the biomedical domain. Unlike other languages, Arabic suffers from lack of resources. This work presen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
20
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 40 publications
(20 citation statements)
references
References 16 publications
0
20
0
Order By: Relevance
“…In a related study, Boudjellal [22] presented a BERTbased model to identify biomedical named entities in Arabic text data that investigates the effectiveness of pertaining a monolingual Bidirectional Encoder Representations from Transformers (BERT) model with a small-scale biomedical dataset on enhancing the model understanding of Arabic biomedical text. When the model's performance was compared to that of two state-of-the-art models, it outperformed both with an F1 score of 85%.…”
Section: Related Workmentioning
confidence: 99%
“…In a related study, Boudjellal [22] presented a BERTbased model to identify biomedical named entities in Arabic text data that investigates the effectiveness of pertaining a monolingual Bidirectional Encoder Representations from Transformers (BERT) model with a small-scale biomedical dataset on enhancing the model understanding of Arabic biomedical text. When the model's performance was compared to that of two state-of-the-art models, it outperformed both with an F1 score of 85%.…”
Section: Related Workmentioning
confidence: 99%
“…The CNN model outperformed other models and achieved an accuracy of 81%. Moreover, recently suggested BERT-based Arabic named entity recognition models [67,68] could also be applied for HS recognition if trained on offensive and derogatory terms. Unlike the previous works, in this paper we propose a model trained on many DA offensive and hate speech datasets as well as MSA and evaluate a new pre-trained model to classify DA.…”
Section: Hate and Offensive Speech Detection In Arabicmentioning
confidence: 99%
“…ProteinLM [269] was recently proposed, which trained a large-scale pre-train model for evolutionary-scale protein sequences, and the trained model is available at 23 . More recently, DeepMind develops Alphafold2 [97] that could predict protein structures with a high accuracy, in the challenging 14th Critical Assessment of protein Structure Prediction (CASP14).…”
Section: Language Models For Biological Sequencesmentioning
confidence: 99%
“…Instead of treating the BioNER as the sequence labeling problem, Sun et al [229] proposed to consider the BioNER as the machine reading comprehension (MRC) problem based on BERT. Besides English, there is much work exploring the pre-trained language models on the BioNER of other languages, including Chinese [43,91,129,130,252,272], Spanish [6,75,159], French [41], Korean [109], Russian [154], Arabic [23], Italian [28]. In Table 6, we summary the commonly used datasets in the BioNER task.…”
Section: Named Entity Recognition Biomedical Named Entity Recognition...mentioning
confidence: 99%