2021
DOI: 10.1038/s43856-021-00008-0
|View full text |Cite
|
Sign up to set email alerts
|

A BERT model generates diagnostically relevant semantic embeddings from pathology synopses with active learning

Abstract: Background Pathology synopses consist of semi-structured or unstructured text summarizing visual information by observing human tissue. Experts write and interpret these synopses with high domain-specific knowledge to extract tissue semantics and formulate a diagnosis in the context of ancillary testing and clinical information. The limited number of specialists available to interpret pathology synopses restricts the utility of the inherent information. Deep learning offers a tool for informati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 11 publications
(15 citation statements)
references
References 41 publications
0
15
0
Order By: Relevance
“…In terms of the weaknesses of our system, WSI labels came from our previously published BERT model's predictions [46] rather than experts, which are not perfectly correct. This was implemented as it is not practically feasible to manually label many hundreds of WSI with keywords from semi-structured diagnostic synopses.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…In terms of the weaknesses of our system, WSI labels came from our previously published BERT model's predictions [46] rather than experts, which are not perfectly correct. This was implemented as it is not practically feasible to manually label many hundreds of WSI with keywords from semi-structured diagnostic synopses.…”
Section: Discussionmentioning
confidence: 99%
“…Labels were created by simplifying the predictions of a fine-tuned BERT model on WSI synopses, as previously reported by our group [46]. This BERT model’s predictions took the form of a multi-label task.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…When the data is textual, the extraction of features is a bit different where the aim is to create word or text embeddings. Generally in medical NLP and at the feature extraction level, the BERT (as a state-of-the-art model) was used in several studies as in [14] , [15] , [16] , [17] , whereas, the Word2Vec model was used in [18] , [19] , [20] . In contrast, and at the algorithmic level, various deep learning models were widely used in medical NLP.…”
Section: Literature Reviewmentioning
confidence: 99%
“…When the data is textual, the extraction of features is a bit different where the aim is to create word or text embeddings. Generally in medical NLP and at the feature extraction level, the BERT (as a state-of-the-art model) was used in several studies as in [14,15,16,17], whereas, the Word2Vec model was used in [18,19,20]. In contrast, and at the algorithmic level, various deep learning models were widely used in medical NLP.…”
Section: Literature Reviewmentioning
confidence: 99%