Data Analytics in Biomedical Engineering and Healthcare 2021
DOI: 10.1016/b978-0-12-819314-3.00005-7
|View full text |Cite
|
Sign up to set email alerts
|

Utilizing BERT for biomedical and clinical text mining

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(8 citation statements)
references
References 26 publications
0
7
0
1
Order By: Relevance
“…Named entity recognition models based on deep learning can recognize neurologic signs and symptoms in the biomedical literature and electronic health records (Table 1). Previous work has shown that BERT outperforms CNNs on recognizing drugs and diseases in annotated test corpora (52,55). We Text spans that identified clinical concepts were longer in the EHR corpus and shortest in the OMIM corpus (see blue inset histograms).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Named entity recognition models based on deep learning can recognize neurologic signs and symptoms in the biomedical literature and electronic health records (Table 1). Previous work has shown that BERT outperforms CNNs on recognizing drugs and diseases in annotated test corpora (52,55). We Text spans that identified clinical concepts were longer in the EHR corpus and shortest in the OMIM corpus (see blue inset histograms).…”
Section: Discussionmentioning
confidence: 99%
“…Although considerable work has been done on automated concept identification of drugs and diseases, less work has been done on the automated identification of signs and symptoms (52). Identifying signs and symptoms is critical to precision medicine and deep phenotyping (56).…”
Section: A Proposed Approachmentioning
confidence: 99%
“…14 This was a significant leap forward in state-of-the-art deep learning-based natural language processing as it allowed for the development of large foundational models (eg, Bidirectional Encoder Representations from Transformers [BERT] 15 and Generative Pretrained Transformer 3 [GPT-3]), 16 pretrained on massive amounts of text data from the Internet. There are now several publicly available Transformer-based language models adapted to the clinical domain through further training on biomedical corpora (eg, ClinicalBERT, 17 , 18 GatorTtron, 19 Clincal-Longformer 20 ). A frequently cited obstacle to creating AI systems is access to labeled data.…”
Section: Proposed Alternativementioning
confidence: 99%
“…Data used by NLP models is pre-processed using techniques such as lemmatisation and stemming – reducing words to a single root for uniformity, keyword extraction – extracting important phrases and keywords, named entity recognition – classification of named entities into predefined categories, and tokenisation – breaking down text into smaller chunks. NLP algorithms perform language processing tasks such as natural language inference, entity extraction, relation extraction, text classification, question answering and text summarization [12] . Such tasks are explained further in Section 3 , and are used as performance indicators of the models compared and discussed in this review.…”
Section: Introductionmentioning
confidence: 99%