2023
DOI: 10.1007/s00330-022-09315-z
|View full text |Cite
|
Sign up to set email alerts
|

Multicentre external validation of a commercial artificial intelligence software to analyse chest radiographs in health screening environments with low disease prevalence

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(7 citation statements)
references
References 30 publications
0
7
0
Order By: Relevance
“…A recent study performed an external validation of an AI algorithm capable of classifying chest X-rays normal or abnormal, through a cohort of real images from two primary care centres, and obtained a sensitivity of 0.35 (95% con dence interval (95% CI) 0.25; 0.48) in classifying normal or abnormal X-rays and an AUROC of 0.65 (95% CI 0.63; 0.66). On the other hand, it obtained an AUROC between 0.77-0.99 in several internal validations (31), highlighting that additional training with data from this environment is necessary to improve their performance when using the models in a speci c clinical environment that differs from the training environment. Another study, conducted in primary care centres and in the emergency department of a hospital in the United Kingdom, concluded that the algorithm could accurately and consistently differentiate abnormalities on an image, obtaining a sensitivity between 0.55 and 0.93, and an AUC between 0.88 and 0.99 depending on the condition, detecting 10 of the most prevalent conditions (32).…”
Section: Introductionmentioning
confidence: 99%
“…A recent study performed an external validation of an AI algorithm capable of classifying chest X-rays normal or abnormal, through a cohort of real images from two primary care centres, and obtained a sensitivity of 0.35 (95% con dence interval (95% CI) 0.25; 0.48) in classifying normal or abnormal X-rays and an AUROC of 0.65 (95% CI 0.63; 0.66). On the other hand, it obtained an AUROC between 0.77-0.99 in several internal validations (31), highlighting that additional training with data from this environment is necessary to improve their performance when using the models in a speci c clinical environment that differs from the training environment. Another study, conducted in primary care centres and in the emergency department of a hospital in the United Kingdom, concluded that the algorithm could accurately and consistently differentiate abnormalities on an image, obtaining a sensitivity between 0.55 and 0.93, and an AUC between 0.88 and 0.99 depending on the condition, detecting 10 of the most prevalent conditions (32).…”
Section: Introductionmentioning
confidence: 99%
“…6 This finding is further reinforced by recent studies that demonstrated commercially available AI systems that, although backed by excellent results in peer-reviewed publications, significantly underperform in clinical practice. Kim et al 7 show that radiologist diagnosis assisted by AI, still shows unsatisfactory improvements compared to radiologist diagnosis without AI, as well as severely increased reading times due to the false readings of the AI. In the work of Wynants et al, 8 several research flaws were discovered in publicly available Covid-19 prediction models and the concern for investigation of possible bias in AI is raised.…”
Section: Introductionmentioning
confidence: 99%
“…A recent study conducting external validation of an AI algorithm designed to classify chest X-rays as normal or abnormal using a cohort of real images from two primary care centres, indicated that additional training with data from these environments was required to enhance the algorithm’s performance, particularly in clinical setting different from its initial training environment 27 .…”
Section: Introductionmentioning
confidence: 99%