Background Racial bias is a key concern regarding the development, validation, and implementation of machine learning (ML) models in clinical settings. Despite the potential of bias to propagate health disparities, racial bias in clinical ML has yet to be thoroughly examined and best practices for bias mitigation remain unclear. Objective Our objective was to perform a scoping review to characterize the methods by which the racial bias of ML has been assessed and describe strategies that may be used to enhance algorithmic fairness in clinical ML. Methods A scoping review was conducted in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-analyses (PRISMA) Extension for Scoping Reviews. A literature search using PubMed, Scopus, and Embase databases, as well as Google Scholar, identified 635 records, of which 12 studies were included. Results Applications of ML were varied and involved diagnosis, outcome prediction, and clinical score prediction performed on data sets including images, diagnostic studies, clinical text, and clinical variables. Of the 12 studies, 1 (8%) described a model in routine clinical use, 2 (17%) examined prospectively validated clinical models, and the remaining 9 (75%) described internally validated models. In addition, 8 (67%) studies concluded that racial bias was present, 2 (17%) concluded that it was not, and 2 (17%) assessed the implementation of bias mitigation strategies without comparison to a baseline model. Fairness metrics used to assess algorithmic racial bias were inconsistent. The most commonly observed metrics were equal opportunity difference (5/12, 42%), accuracy (4/12, 25%), and disparate impact (2/12, 17%). All 8 (67%) studies that implemented methods for mitigation of racial bias successfully increased fairness, as measured by the authors’ chosen metrics. Preprocessing methods of bias mitigation were most commonly used across all studies that implemented them. Conclusions The broad scope of medical ML applications and potential patient harms demand an increased emphasis on evaluation and mitigation of racial bias in clinical ML. However, the adoption of algorithmic fairness principles in medicine remains inconsistent and is limited by poor data availability and ML model reporting. We recommend that researchers and journal editors emphasize standardized reporting and data availability in medical ML studies to improve transparency and facilitate evaluation for racial bias.
We show here that computer game players can build high-quality crystal structures. Introduction of a new feature into the computer game Foldit allows players to build and real-space refine structures into electron density maps. To assess the usefulness of this feature, we held a crystallographic model-building competition between trained crystallographers, undergraduate students, Foldit players and automatic model-building algorithms. After removal of disordered residues, a team of Foldit players achieved the most accurate structure. Analysing the target protein of the competition, YPL067C, uncovered a new family of histidine triad proteins apparently involved in the prevention of amyloid toxicity. From this study, we conclude that crystallographers can utilize crowdsourcing to interpret electron density information and to produce structure solutions of the highest quality.
Background This study aimed to build and evaluate a deep learning, artificial intelligence (AI) model to automatically classify swallow types based on raw data from esophageal high‐resolution manometry (HRM). Methods HRM studies on patients with no history of esophageal surgery were collected including 1,741 studies with 26,115 swallows labeled by swallow type (normal, hypercontractile, weak‐fragmented, failed, and premature) by an expert interpreter per the Chicago Classification. The dataset was stratified and split into train/validation/test datasets for model development. Long short‐term memory (LSTM), a type of deep‐learning AI model, was trained and evaluated. The overall performance and detailed per‐swallow type performance were analyzed. The interpretations of the supine swallows in a single study were further used to generate an overall classification of peristalsis. Key Results The LSTM model for swallow type yielded accuracies from the train/validation/test datasets of 0.86/0.81/0.83. The model's interpretation for study‐level classification of peristalsis yielded accuracy of 0.88 in the test dataset. Among model misclassification, 535/698 (77%) swallows and 25/35 (71%) studies were to adjacent categories, for example, normal to weak or normal to ineffective, respectively. Conclusions and Inferences A deep‐learning AI model can automatically and accurately identify the Chicago Classification swallow types and peristalsis classification from raw HRM data. While future work to refine this model and incorporate overall manometric diagnoses are needed, this study demonstrates the role that AI will serve in the interpretation and classification of esophageal HRM studies.
BACKGROUND Racial bias is a key concern regarding the development, validation, and implementation of machine learning (ML) models in clinical settings. Despite the potential of bias to propagate health disparities, racial bias in clinical ML has yet to be thoroughly examined and best practices for bias mitigation remain unclear. OBJECTIVE Our objective was to perform a scoping review to characterize the methods by which racial bias of ML has been assessed and describe strategies that may be used to enhance algorithmic fairness in clinical ML. METHODS A scoping review was conducted in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-analyses Extension for Scoping Reviews. Literature search using the PubMed, Scopus, and Embase databases as well as Google Scholar identified 635 records, of which 12 studies were included. RESULTS Applications of ML were varied and involved diagnosis, outcome prediction, and clinical score prediction performed on datasets including images, diagnostic studies, clinical text, and clinical variables. One study (8%) described a model in routine clinical use, two (17%) examined prospectively validated clinical models, and the remaining nine (75%) described internally validated models. Eight studies (75%) concluded that racial bias was present, two (17%) concluded that it was not, and two (17%) assessed the implementation of bias mitigation strategies without comparison to a baseline model. Fairness metrics used to assess algorithmic racial bias were inconsistent. The most commonly observed metrics were: equal opportunity difference (5/12, 42%); accuracy (4/12, 25%); and disparate impact (2/12, 17%). All eight studies (67%) which implemented methods for mitigation of racial bias successfully increased fairness as measured by the authors’ chosen metrics. Pre-processing methods of bias mitigation were the most commonly used across all studies which implemented them. CONCLUSIONS The broad scope of medical ML applications and potential patient harms demand an increased emphasis on evaluation and mitigation of racial bias in clinical ML. However, the adoption of algorithmic fairness principles in medicine remains inconsistent and is limited by poor data availability and ML model reporting. We recommend that researchers and journal editors emphasize standardized reporting and data availability in medical ML studies to improve transparency and facilitate evaluation for racial bias.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.