2021
DOI: 10.48550/arxiv.2104.09585
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ELECTRAMed: a new pre-trained language representation model for biomedical NLP

Giacomo Miolo,
Giulio Mantoan,
Carlotta Orsenigo

Abstract: The overwhelming amount of biomedical scientific texts calls for the development of effective language models able to tackle a wide range of biomedical natural language processing (NLP) tasks. The most recent dominant approaches are domain-specific models, initialized with general-domain textual data and then trained on a variety of scientific corpora. However, it has been observed that for specialized domains in which large corpora exist, training a model from scratch with just in-domain knowledge may yield b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 17 publications
0
3
0
Order By: Relevance
“…Consequently, they propose PubMedBERT which is pre-trained on 14M PubMed abstracts from scratch. Similarly, [20] pre-train on 28M data as in [8] also from scratch, using the more advanced ELECTRA model. All these works have shown improved performance on plenty of biomedical literature language processing tasks compared to the original BERT pretrained on general domain, while none of them is for biomedical generation tasks.…”
Section: Pre-trained Language Models In Biomedical Domainmentioning
confidence: 99%
“…Consequently, they propose PubMedBERT which is pre-trained on 14M PubMed abstracts from scratch. Similarly, [20] pre-train on 28M data as in [8] also from scratch, using the more advanced ELECTRA model. All these works have shown improved performance on plenty of biomedical literature language processing tasks compared to the original BERT pretrained on general domain, while none of them is for biomedical generation tasks.…”
Section: Pre-trained Language Models In Biomedical Domainmentioning
confidence: 99%
“…The authors of the papers [5][6][7] involve architectures based on sequential and transformer networks to solve analytical medical problems, which can be applied to our task for the transfer of knowledge in the cross-domain field.…”
Section: Literature Reviewmentioning
confidence: 99%
“…(3) Pre-training language models are an open research area that guide the acquisition of striking gains on several natural language processing tasks. Some of the language representation models such as BERT (Devlin, Chang, Lee, & Toutanova, 2018) and ELMO have been trained generally for general domain datasets, and they perform weakly on domain-specific text mining tasks (Lee et al, 2020), thereby leading to the development of domain-specific Khadivi and Sato: A Bibliometric Study of NLP language models like Bio-BERT and ELECTRAMed (Miolo, Mantoan, & Orsenigo, 2021), which are applicable for biomedical text mining, or BERT_SE (Fávero & Casanova, 2021), which is used for noticing software engineering vocabulary, which was initiated a few years ago. Additionally, in recent years, BERT has been used for fake news detection.…”
Section: Discovering Research Trends and Future Directionmentioning
confidence: 99%