2020
DOI: 10.1007/978-3-030-45442-5_35
|View full text |Cite
|
Sign up to set email alerts
|

On Biomedical Named Entity Recognition: Experiments in Interlingual Transfer for Clinical and Social Media Texts

Abstract: Although deep neural networks yield state-of-the-art performance in biomedical named entity recognition (bioNER), much research shares one limitation: models are usually trained and evaluated on English texts from a single domain. In this work, we present a fine-grained evaluation intended to understand the efficiency of multilingual BERT-based models for bioNER of drug and disease mentions across two domains in two languages, namely clinical data and user-generated texts on drug therapy in English and Russian… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
10
1
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 16 publications
(12 citation statements)
references
References 18 publications
0
10
1
1
Order By: Relevance
“…Instead of treating the BioNER as the sequence labeling problem, Sun et al [229] proposed to consider the BioNER as the machine reading comprehension (MRC) problem based on BERT. Besides English, there is much work exploring the pre-trained language models on the BioNER of other languages, including Chinese [43,91,129,130,252,272], Spanish [6,75,159], French [41], Korean [109], Russian [154], Arabic [23], Italian [28]. In Table 6, we summary the commonly used datasets in the BioNER task.…”
Section: Named Entity Recognition Biomedical Named Entity Recognition...mentioning
confidence: 99%
“…Instead of treating the BioNER as the sequence labeling problem, Sun et al [229] proposed to consider the BioNER as the machine reading comprehension (MRC) problem based on BERT. Besides English, there is much work exploring the pre-trained language models on the BioNER of other languages, including Chinese [43,91,129,130,252,272], Spanish [6,75,159], French [41], Korean [109], Russian [154], Arabic [23], Italian [28]. In Table 6, we summary the commonly used datasets in the BioNER task.…”
Section: Named Entity Recognition Biomedical Named Entity Recognition...mentioning
confidence: 99%
“…Following our previous work on NER (Miftahutdinov et al, 2020), we utilize different BERT models with a softmax layer over all possible tags as the output for NER. Word labels are encoded with the BIO tag scheme.…”
Section: Pre-training and Fine-tuning Domain-specific Bertmentioning
confidence: 99%
“…NER is applied to various domains, and the set of entities is selected based on the given domain, for example, NER has been used to extract drug and disease mentions in English and Russian by Miftahutdinov and others [1]. In addition, Chen and others [2] investigated clinical NER tasks and compared different active learning algorithms for NER, and Carbonell and others [3] applied NER to handwritten data jointly with text localization and transcription.…”
Section: Introductionmentioning
confidence: 99%
“…These results represent state-of-the-art performance on the Persian NER task.Named entity recognition (NER) detects and classifies named entities in text into categories, for example, persons, organizations, locations, and time expressions. NER is a fundamental task in many natural language processing (NLP) applications, for example, information extraction, text summarization, and machine translation, especially for low-resource languages, for example, Persian, which requires additional features to process the text effectively.NER is applied to various domains, and the set of entities is selected based on the given domain, for example, NER has been used to extract drug and disease mentions in English and Russian by Miftahutdinov and others [1]. In addition, Chen and others [2] investigated…”
mentioning
confidence: 99%