2021
DOI: 10.1016/j.compbiolchem.2021.107529
|View full text |Cite
|
Sign up to set email alerts
|

Ensembling machine learning models to boost molecular affinity prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

3
4

Authors

Journals

citations
Cited by 15 publications
(13 citation statements)
references
References 57 publications
0
13
0
Order By: Relevance
“…This ensemble approach has been shown multiple times to improve predictions of machine learning and deep learning models (Hansen and Salamon 1990;Ashtawy and Mahapatra 2015;Ericksen et al, 2017;Francoeur et al, 2020;Kwon et al, 2020;Meli et al, 2021). More generally, a consensus score amongst multiple models (also with different architectures) can be used as well (Druchok et al, 2021), and the average between different models (different architectures and/or different training data sets) has been shown to improve pose predictions with CNN scoring functions (McNutt et al, 2021). While the average across different models is often used to estimate the performance of the ensemble, the standard deviation across predictions gives information about their stability and can be used as a diagnostic tool.…”
Section: Discussionmentioning
confidence: 99%
“…This ensemble approach has been shown multiple times to improve predictions of machine learning and deep learning models (Hansen and Salamon 1990;Ashtawy and Mahapatra 2015;Ericksen et al, 2017;Francoeur et al, 2020;Kwon et al, 2020;Meli et al, 2021). More generally, a consensus score amongst multiple models (also with different architectures) can be used as well (Druchok et al, 2021), and the average between different models (different architectures and/or different training data sets) has been shown to improve pose predictions with CNN scoring functions (McNutt et al, 2021). While the average across different models is often used to estimate the performance of the ensemble, the standard deviation across predictions gives information about their stability and can be used as a diagnostic tool.…”
Section: Discussionmentioning
confidence: 99%
“…However, the proposed method successfully avoids the overfitting, which can be deduced from the fact that the trained model gives accurate capacity estimation for cell #8 despite the training data includes abnormal capacity drop of cell #5. This is also beneficial from the inherent voting mechanism under ensembling framework, which makes the trained model prefer the majority rather than focusing on the outlier [43].…”
Section: E More Discussionmentioning
confidence: 99%
“…Another two‐stage ensembling pipeline is suggested in Ref. [23], uniting six ML techniques to enhance the DTA prediction. The methods include Support Vector Machine, Random Forest, CatBoost, 24 feed‐forward neural network, graph neural network, and Bidirectional Encoder Representations from Transformers 25 …”
Section: Introductionmentioning
confidence: 99%