2005
DOI: 10.1016/j.chemolab.2004.11.001
|View full text |Cite
|
Sign up to set email alerts
|

A performance comparison of modern statistical techniques for molecular descriptor selection and retention prediction in chromatographic QSRR studies

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
60
0

Year Published

2006
2006
2017
2017

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 131 publications
(60 citation statements)
references
References 26 publications
0
60
0
Order By: Relevance
“…Thus, this usually works to decrease undesirable effects of illdistribution of samples. It is reported that, compared with boosting, bagging has a superior advantage to have a robustness against noises [13,31]. On the other hand, boosting is based on the idea of intensive learning of 'bad' samples.…”
Section: Prediction Performances Of the Nir Tobacco Datamentioning
confidence: 98%
See 1 more Smart Citation
“…Thus, this usually works to decrease undesirable effects of illdistribution of samples. It is reported that, compared with boosting, bagging has a superior advantage to have a robustness against noises [13,31]. On the other hand, boosting is based on the idea of intensive learning of 'bad' samples.…”
Section: Prediction Performances Of the Nir Tobacco Datamentioning
confidence: 98%
“…The main advantage of such ensemble techniques is that they increase the accuracy and stability of the predictor by combining several hypotheses. Such ensemble learning methods have recently been introduced into the field of chemometrics [12,13,15,16]. He et al [12] showed that the boosting tree improves the prediction performance of any single classification method in the classification for chemical data.…”
Section: Introductionmentioning
confidence: 99%
“…This randomness has the effect of building new trees with different structures, increasing the variety of relationships modelled within the forest, which in turn improves the overall predictive performance and makes it robust against overfitting. However, while RF are good for identifying main trends, it can lead to some points being poorly predicted (Hancock et al, 2005).…”
Section: Introductionmentioning
confidence: 99%
“…By using many trees, RF have the ability to identify important variables and their relationship with the response. The number of trees, therefore, needs to be set sufficiently high to allow for this convergence as well as to identify the most important variables (Goudarzi and Shahsavani, 2012;Hancock et al, 2005). This randomness has the effect of building new trees with different structures, increasing the variety of relationships modelled within the forest, which in turn improves the overall predictive performance and makes it robust against overfitting.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation