2019
DOI: 10.1021/acs.jmedchem.9b01101
|View full text |Cite
|
Sign up to set email alerts
|

Interpretation of Compound Activity Predictions from Complex Machine Learning Models Using Local Approximations and Shapley Values

Abstract: In qualitative or quantitative studies of structure−activity relationships (SARs), machine learning (ML) models are trained to recognize structural patterns that differentiate between active and inactive compounds. Understanding model decisions is challenging but of critical importance to guide compound design. Moreover, the interpretation of ML results provides an additional level of model validation based on expert knowledge. A number of complex ML approaches, especially deep learning (DL) architectures, hav… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
198
0
2

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 264 publications
(201 citation statements)
references
References 61 publications
1
198
0
2
Order By: Relevance
“…The SHapley Additive exPlanations (SHAP) approach is an extension of LIME; feature weights are represented as SHapley values from game theory. The SHAP approach has a high potential for rationalizing the predictions made by complex ML models (37). In the present study, we used the SHAP method to observe the influence of each feature on the prediction results during the prediction process applied to each sample.…”
Section: Model Interpretationmentioning
confidence: 99%
“…The SHapley Additive exPlanations (SHAP) approach is an extension of LIME; feature weights are represented as SHapley values from game theory. The SHAP approach has a high potential for rationalizing the predictions made by complex ML models (37). In the present study, we used the SHAP method to observe the influence of each feature on the prediction results during the prediction process applied to each sample.…”
Section: Model Interpretationmentioning
confidence: 99%
“…Pope et al 67 adapted gradient-based feature attribution 68,69 for the identification of relevant functional groups in adverse effect prediction 70 . Recently, SHAP 52 was used to interpret relevant features for compound potency and multitarget activity prediction 71 . Hochuli et al 72 compared several feature attribution methodologies, showing how the visualization of attributions assists in the parsing and interpretation of protein-ligand scoring with three-dimensional convolutional neural networks 73,74 .…”
Section: Relevance Of Input Featuresmentioning
confidence: 99%
“…In the context of activity predictions, Shapley values can also be rationalized as a fair or reasonable allocation of feature importance given a particular model output [19]. Features contribute to the model's output or prediction with different magnitude and sign, which is accounted for by Shapley values.…”
Section: Shapley Valuesmentioning
confidence: 99%