2021
DOI: 10.1093/jamia/ocab126
|View full text |Cite
|
Sign up to set email alerts
|

MT-clinical BERT: scaling clinical information extraction with multitask learning

Abstract: Objective Clinical notes contain an abundance of important, but not-readily accessible, information about patients. Systems that automatically extract this information rely on large amounts of training data of which there exists limited resources to create. Furthermore, they are developed disjointly, meaning that no information can be shared among task-specific systems. This bottleneck unnecessarily complicates practical application, reduces the performance capabilities of each individual sol… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 29 publications
(7 citation statements)
references
References 12 publications
0
7
0
Order By: Relevance
“…Similarly, there have been various attempts to develop domain-specific LMs to extract key information from the respective custom text data. Similarly, clinicalBERT is trained on 2 million clinical notes in the MIMIC-III v1.4 database; BioBERT is a biomedical-specific language model; mBERT is developed for multilingual machine translation tasks; patentBERT is built for patent classification; and FinBERT is built for financial NLP tasks . Some of the recently built LMs that have specific applications to process engineering include Ex-SciBERT, which deals with hydrogen production from alcohol-related literature for sentence classification, and NER; process-BERT was built for process informatics; and Battery-BERT was developed on battery-related scientific articles.…”
Section: Background and Contributionmentioning
confidence: 99%
“…Similarly, there have been various attempts to develop domain-specific LMs to extract key information from the respective custom text data. Similarly, clinicalBERT is trained on 2 million clinical notes in the MIMIC-III v1.4 database; BioBERT is a biomedical-specific language model; mBERT is developed for multilingual machine translation tasks; patentBERT is built for patent classification; and FinBERT is built for financial NLP tasks . Some of the recently built LMs that have specific applications to process engineering include Ex-SciBERT, which deals with hydrogen production from alcohol-related literature for sentence classification, and NER; process-BERT was built for process informatics; and Battery-BERT was developed on battery-related scientific articles.…”
Section: Background and Contributionmentioning
confidence: 99%
“…In the clinical domain, Biseda et al [33] predict ICD codes from hospital notes by utilizing the Clinical BERT variant. Mulyar et al [34] developed Multitask-Clinical BERT for multitasking information extraction.…”
Section: Previous Workmentioning
confidence: 99%
“…These example applications focus on general corpora from Wikipedia and news abstracts, respectively. In the medical domain, language models were used for multi-task information extraction from medical notes [24]. They were also used for feature extraction from Twitter posts to identify patients at risk of developing depression [25].…”
Section: Related Workmentioning
confidence: 99%