2020
DOI: 10.1007/978-3-030-61166-8_27
|View full text |Cite
|
Sign up to set email alerts
|

Labelling Imaging Datasets on the Basis of Neuroradiology Reports: A Validation Study

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

3
4

Authors

Journals

citations
Cited by 10 publications
(17 citation statements)
references
References 10 publications
0
17
0
Order By: Relevance
“…Previous studies have only reported model performance on a hold-out set of labelled reports [6,7,9], and to date, there has been no investigation into the general validity of NLP-derived labels for head MRI examinations [36]. An important question Table 2 Reference-standard report labels across all abnormality categories.…”
Section: Nlp Modellingmentioning
confidence: 99%
“…Previous studies have only reported model performance on a hold-out set of labelled reports [6,7,9], and to date, there has been no investigation into the general validity of NLP-derived labels for head MRI examinations [36]. An important question Table 2 Reference-standard report labels across all abnormality categories.…”
Section: Nlp Modellingmentioning
confidence: 99%
“…Creating a manual annotation protocol is difficult [6] and the protocol constantly evolves as new data are encountered and labelled. It is therefore useful to be able to encode certain phrases/rules from the protocol in a template so that they can be learned by the model.…”
Section: Protocol-derived Templatesmentioning
confidence: 99%
“…However, extracting labels from text can be challenging because the language in radiology reports is diverse, domain-specific, and often difficult to interpret. Therefore, the task of reading the radiology report and assigning labels is not trivial and requires a certain degree of medical knowledge on the part of a human annotator [6]. When we rely on pure data-driven learning, we find that the model sometimes fails to learn critical features or learns the correct answer via simple heuristics (e.g., that presence of the word "likely" indicates positivity) rather than valid reasoning, and thus fails to generalise to rarer cases which have not been learned or where the heuristics break down (e.g., "likely represents prominent VR space or lacunar infarct" which indicates uncertainty over two differential diagnoses).…”
Section: Introductionmentioning
confidence: 99%
“…Generating these labels can be performed manually (each report is read by a human rater who assigns diagnostic codes or other labels) or automatically (a rules-based or machine-learningbased NLP technique automatically assigns labels to reports). The latter techniques usually rely on a subset of reports which have been annotated manually to allow supervised learning [4][5][6]. Labels generated manually are often referred to as the "ground truth".…”
Section: Background and Significancementioning
confidence: 99%