2002
DOI: 10.1214/aos/1015362183
|View full text |Cite
|
Sign up to set email alerts
|

Empirical Margin Distributions and Bounding the Generalization Error of Combined Classifiers

Abstract: We prove new probabilistic upper bounds on generalization error of complex classifiers that are combinations of simple classifiers. Such combinations could be implemented by neural networks or by voting methods of combining the classifiers, such as boosting and bagging. The bounds are in terms of the empirical distribution of the margin of the combined classifier. They are based on the methods of the theory of Gaussian and empirical processes (comparison inequalities, symmetrization method, concentration inequ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
253
0
1

Year Published

2005
2005
2020
2020

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 310 publications
(259 citation statements)
references
References 31 publications
5
253
0
1
Order By: Relevance
“…Together with Theorem 3.2, this result establishes a precise characterization of the computational and statistical behavior of AdaBoost for all iterations above the threshold T (n), and notably complements the classical margin upper bounds [81,53]. Thus, Theorem 3.4 reinforces a crucial conclusion from Section 3.1-the max-1 -margin is the key quantity governing the generalization error of AdaBoost, and as emphasized earlier, this resolves a series of discussions around this topic [81,15].…”
Section: Boosting In High Dimensionssupporting
confidence: 66%
See 3 more Smart Citations
“…Together with Theorem 3.2, this result establishes a precise characterization of the computational and statistical behavior of AdaBoost for all iterations above the threshold T (n), and notably complements the classical margin upper bounds [81,53]. Thus, Theorem 3.4 reinforces a crucial conclusion from Section 3.1-the max-1 -margin is the key quantity governing the generalization error of AdaBoost, and as emphasized earlier, this resolves a series of discussions around this topic [81,15].…”
Section: Boosting In High Dimensionssupporting
confidence: 66%
“…• Turning to boosting, we provide a sharp characterization of a threshold T such that for all iterations t ≥ T , the AdaBoost iterates (with a properly scaled step size) stay arbitrarily close toθ n, 1 , in the large n, p limit (1.1) (Theorem 3.4). Together with Theorems 3.1-3.2, this result immediately provides an exact characterization of the generalization error of boosting, and improves upon the existing upper bounds [81,53] by a margin. This formula crucially involves κ , and therefore, our results resolve an open question posed in Schapire et al [81], Breiman [15] regarding which quantity truly governs the generalization performance of AdaBoost.…”
Section: Introductionsupporting
confidence: 71%
See 2 more Smart Citations
“…This section analyzes the generalization error of the proposed CoUDA method based on Rademacher complexity [24], [56], [34], [57]. Before that, we give some necessary notations.…”
Section: Theoretical Analysismentioning
confidence: 99%