2019
DOI: 10.1021/acs.jcim.9b00297
|View full text |Cite
|
Sign up to set email alerts
|

Reliable Prediction Errors for Deep Neural Networks Using Test-Time Dropout

Abstract: While the use of deep learning in drug discovery is gaining increasing attention, the lack of methods to computate reliable errors in prediction for Neural Networks prevents their application to guide decision making in domains where identifying unreliable predictions is essential, e.g. precision medicine. Here, we present a framework to compute reliable errors in prediction for Neural Networks using Test-Time Dropout and Conformal Prediction. Specifically, the algorithm consists of training a single Neural Ne… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
69
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 51 publications
(73 citation statements)
references
References 54 publications
4
69
0
Order By: Relevance
“…The scaled pIC50 MSE values in Table 1 translate to a root mean squared error of approximately 0.8 pIC50 units. These values are in the range of expected errors for ML models trained on heterogeneous ChEMBL25 data (Kalliokoski et al, 2013 ), and are in agreement with prior literature that also demonstrated that RF and FFN models approached the upper limit of overall accuracy across the dataset, given the heterogeneous IC50 measurements in ChEMBL25 (Cortés-Ciriano and Bender, 2019a ).…”
Section: Resultssupporting
confidence: 90%
See 2 more Smart Citations
“…The scaled pIC50 MSE values in Table 1 translate to a root mean squared error of approximately 0.8 pIC50 units. These values are in the range of expected errors for ML models trained on heterogeneous ChEMBL25 data (Kalliokoski et al, 2013 ), and are in agreement with prior literature that also demonstrated that RF and FFN models approached the upper limit of overall accuracy across the dataset, given the heterogeneous IC50 measurements in ChEMBL25 (Cortés-Ciriano and Bender, 2019a ).…”
Section: Resultssupporting
confidence: 90%
“…Interestingly, there is very little variation in the size of the FFN confidence intervals across all predictions on the validation set ( Figure 4A ) but this is still sufficient for the FFN to generate valid prediction intervals ( Figure 4B ). In total, conformal prediction is able to accurately gauge both RF and FFN model confidence for predictions on held-out validation data, in agreement with prior literature (Svensson et al, 2018 ; Cortés-Ciriano and Bender, 2019a ).…”
Section: Resultssupporting
confidence: 86%
See 1 more Smart Citation
“…Further information about the data sets is given in Table 1 and in a previous study by the authors [42]. We also collected 25 QSAR data sets for validation from previous work by the authors (Table 2) [42][43][44]. All data sets used in this study, as well as the code required to generate the results presented herein, are publicly available at https ://githu b.com/isidr oc/QAFFP _regre ssion .…”
Section: Data Collection and Curationmentioning
confidence: 99%
“…Having trained and validated the signaturizers, we massively inferred missing signatures for the ~800,000 molecules available in the CC, obtaining a complete set of 25x128-dimensional signatures for each molecule (chemicalchecker.org/downloads). To explore the reliability of the inferred signatures, we assigned an 'applicability' score (α) to predictions based on the following: (a) the proximity of a predicted signature to true (experimental) signatures available in the training set; (b) the robustness of the SNN output to a test-time data dropout 10 ; and (c) the accuracy expected a priori based on the experimental CC datasets available for the molecule (Figure 2a).…”
Section: Large-scale Inference Of Bioactivity Signaturesmentioning
confidence: 99%