2016
DOI: 10.1016/j.csda.2015.06.010
|View full text |Cite
|
Sign up to set email alerts
|

Noise peeling methods to improve boosting algorithms

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
6
3
1

Relationship

1
9

Authors

Journals

citations
Cited by 12 publications
(5 citation statements)
references
References 42 publications
0
5
0
Order By: Relevance
“…Although bagging and random forests generally perform better than AdaBoost under noisy circumstances, they are still not completely robust to noise and outliers [Maclin and Opitz, 1997]. Deleting outliers (also called noise filtering or noise peeling) by pre-processing the data is preferable under certain high noise circumstances [Martinez and Gray, 2016].…”
Section: Qmm Performance Under Noisementioning
confidence: 99%
“…Although bagging and random forests generally perform better than AdaBoost under noisy circumstances, they are still not completely robust to noise and outliers [Maclin and Opitz, 1997]. Deleting outliers (also called noise filtering or noise peeling) by pre-processing the data is preferable under certain high noise circumstances [Martinez and Gray, 2016].…”
Section: Qmm Performance Under Noisementioning
confidence: 99%
“…Boosting ensemble is another common metal algorithm. The algorithm was developed from the Probably Approximately Correct (PAC) framework coined by Kearns and Valiant in an effort to enhance the accuracy of classifications [59]. Essentially, boosting was developed in an effort to convert the weak learners into strong learners to bolster their overall classification accuracy [44].…”
Section: Boostingmentioning
confidence: 99%
“…The bound in (4) suggests the generalization error would degrade as the number of rounds of boosting T increased. This in fact happens, especially in cases where there is noise, (see, e.g., Grove and Schuurmans 1998;Dietterich 2000;Long and Servedio 2010;Martinez and Gray 2016), but under certain conditions boosting methods have shown not to overfit even when thousands of rounds are run (see, e.g., Breiman 1996b;Drucker and Cortes 1996;Quinlan 1996;Grove and Schuurmans 1998;Opitz and Maclin 1999;Buja 2000;Lugosi and Vayatis 2004). To better explain the effectiveness of AdaBoost and other ensembles, Schapire et al (1998) developed an upper bound on the generalization error of any ensemble method, based on the margins of the training data, from which it was concluded that larger margins should lead to a lower generalization error of the ensemble, everything else being equal.…”
Section: Generalization Error Bounds Based On Marginsmentioning
confidence: 99%