2021
DOI: 10.1148/ryai.2021210097
|View full text |Cite
|
Sign up to set email alerts
|

Toward Generalizability in the Deployment of Artificial Intelligence in Radiology: Role of Computation Stress Testing to Overcome Underspecification

Abstract: T he use of artificial intelligence (AI) techniques is transforming both the clinical and research fields of medical imaging. As highlighted by the literature in Radiology issues from this year, the use of AI in imaging is an active research field. In 2020, 25% of the articles published in Radiology discussed AI and machine learning (1), including several overviews based on research activity and methods (1-4). The frequency of publications related to AI is even more impressive if we consider the recently launc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
37
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 82 publications
(37 citation statements)
references
References 72 publications
0
37
0
Order By: Relevance
“… 16 Lastly, model development will need to implement stress testing to address the issue of underspecification, or failing to capture the inner logic of an underlying system due to confounding factors in data distributions. 27 To ensure broad generalizability, these stress tests should become standard practice, just as crash tests are fundamental to the automotive industry.…”
Section: Discussionmentioning
confidence: 99%
“… 16 Lastly, model development will need to implement stress testing to address the issue of underspecification, or failing to capture the inner logic of an underlying system due to confounding factors in data distributions. 27 To ensure broad generalizability, these stress tests should become standard practice, just as crash tests are fundamental to the automotive industry.…”
Section: Discussionmentioning
confidence: 99%
“…Medical imaging research is facing a proliferation of studies proposing new AI-powered diagnostic and predictive tools, and promising outstanding performances. Relatively few studies have focused on evaluating the applicability of these models, and what concrete benefits they would bring to clinical practice in the real-world setting [ 110 , 111 , 112 ]: the results can be surprisingly disappointing, when such models are tested in different cohorts of patients or institutions. In the emergency setting, for example, disappointing results achieved by some software should warn radiologists against placing excessive reliance on these tools 90 .…”
Section: Challenges and Perspectivesmentioning
confidence: 99%
“…The reasons behind the translational gap in real-world clinical practice is multi-factorial, partially explained by technical and infrastructure hurdles, lack of IT resources, and no clear data-driven clinical utility analyses. Many of these barriers to adoption are being addressed with existing or emerging solutions [7,8,9,10]. Yet even with successful site specific model validation and successful integration/deployment in clinical workflows, a fundamental problem remains: what happens after the AI model goes into production?…”
Section: Introductionmentioning
confidence: 99%