2020
DOI: 10.1186/s13321-020-0417-9
|View full text |Cite
|
Sign up to set email alerts
|

Comparison and improvement of the predictability and interpretability with ensemble learning models in QSPR applications

Abstract: Ensemble learning helps improve machine learning results by combining several models and allows the production of better predictive performance compared to a single model. It also benefits and accelerates the researches in quantitative structure-activity relationship (QSAR) and quantitative structure-property relationship (QSPR). With the growing number of ensemble learning models such as random forest, the effectiveness of QSAR/QSPR will be limited by the machine's inability to interpret the predictions to re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
52
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 56 publications
(52 citation statements)
references
References 34 publications
0
52
0
Order By: Relevance
“…To study the effect of training with a different tier-one algorithm (meta-learner) on the overall results, the decision trees were introduced in the architecture as an alternative to the SVM. Decision trees are a well-known method and are fast to train [ 42 ]. The result from this modified blended ensemble model (ECNN-DT) was also better than the individual base models.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…To study the effect of training with a different tier-one algorithm (meta-learner) on the overall results, the decision trees were introduced in the architecture as an alternative to the SVM. Decision trees are a well-known method and are fast to train [ 42 ]. The result from this modified blended ensemble model (ECNN-DT) was also better than the individual base models.…”
Section: Resultsmentioning
confidence: 99%
“…The base inducers can be combined using methods such as stack generalization/blending [ 39 ], using different algebraic functions [ 40 ], non-linear combination methods (for instance, Dempster–Shafer belief methods) [ 41 ]. Stacked generalization is an ensemble learning approach that applies a meta-learner and out-of-fold prediction of the training set to detect the best way of combining the base models’ outputs [ 42 ]. A variant of stacked generalization is blending.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Overall, a critical and controversial point of chemical descriptors is their interpretability and physical meaning. In predictive models, it is open for discussion if the descriptors do not only show how a good statistical association between the chemical structure and the property (e.g., biological activity) of interest but if the descriptors can actually explain or contribute to the causality of the activity as encoded by the chemical descriptors [ 58 , 59 ].…”
Section: Open Resources To Expand and Describe The Chemical Spacementioning
confidence: 99%
“…The numerical values obtained from computational chemistry and Morgan fingerprinting [ 15 ] were used as the input data. To obtain the interpretability of the machine learning model, regression coefficients were obtained from the linear model, and the feature importance was obtained from the prediction model based on the decision tree [ 16 ]. To verify the interpretability of the model, we attempted to explain the reaction mechanism [ 17 ] of singlet oxygen scavenging from the feature importance.…”
Section: Introductionmentioning
confidence: 99%