2021
DOI: 10.1186/s13195-021-00879-4
|View full text |Cite
|
Sign up to set email alerts
|

Data analysis with Shapley values for automatic subject selection in Alzheimer’s disease data sets using interpretable machine learning

Abstract: Background For the recruitment and monitoring of subjects for therapy studies, it is important to predict whether mild cognitive impaired (MCI) subjects will prospectively develop Alzheimer’s disease (AD). Machine learning (ML) is suitable to improve early AD prediction. The etiology of AD is heterogeneous, which leads to high variability in disease patterns. Further variability originates from multicentric study designs, varying acquisition protocols, and errors in the preprocessing of magneti… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
9

Relationship

1
8

Authors

Journals

citations
Cited by 36 publications
(17 citation statements)
references
References 79 publications
0
17
0
Order By: Relevance
“…We explained the classification prediction of our model with SHAP [ 47 , 48 ], which is one of the methods to explain the model with feature importance and is inspired by the concept of coalition game theory by replacing “player” in coalition game theory with “data feature” in tabular data. In the ML context, the agents correspond to the features of the data, and the goal is to explain the model's prediction.…”
Section: Methodsmentioning
confidence: 99%
“…We explained the classification prediction of our model with SHAP [ 47 , 48 ], which is one of the methods to explain the model with feature importance and is inspired by the concept of coalition game theory by replacing “player” in coalition game theory with “data feature” in tabular data. In the ML context, the agents correspond to the features of the data, and the goal is to explain the model's prediction.…”
Section: Methodsmentioning
confidence: 99%
“…In addition, we used the Shapley Additive Explanation (SHAP) method to apply the post hoc explainability on the prediction models based on CatBoost, GBDT and XGBoost classifiers, in order to interpret the impact of variables on the prediction outcome. SHAP uses game theory for evaluating the impact of specific input variables to the outcome of a certain model ( 36 ). Moreover, decision curve analysis (DCA) was applied to evaluate the net benefit of the prediction models based on CatBoost, GBDT and XGBoost algorithms.…”
Section: Methodsmentioning
confidence: 99%
“…The prediction abilities of various machine learning models were examined based on the area under the curve of receiver operating characteristics (AUCs) and precision-recall curves of each model. As the methods of prediction in machine learning models are often unclear, we used SHapley Additive exPlanation (SHAP) values to provide accurate attribution values for each clinical feature in our prediction model [33][34][35]. The data were analyzed by using Python (Python Software Foundation version 3.7.6, available at http://www.python.org, accessed on 1 November 2021).…”
Section: Machine Learning Algorithm and Statistical Analysismentioning
confidence: 99%