2007
DOI: 10.1109/tpami.2007.250609
|View full text |Cite
|
Sign up to set email alerts
|

A Comparison of Decision Tree Ensemble Creation Techniques

Abstract: We experimentally evaluate bagging and seven other randomization-based approaches to creating an ensemble of decision tree classifiers. Statistical tests were performed on experimental results from 57 publicly available data sets. When cross-validation comparisons were tested for statistical significance, the best method was statistically more accurate than bagging on only eight of the 57 data sets. Alternatively, examining the average ranks of the algorithms across the group of data sets, we find that boostin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

1
207
1
3

Year Published

2012
2012
2023
2023

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 386 publications
(212 citation statements)
references
References 12 publications
1
207
1
3
Order By: Relevance
“…In turn, neural networks are less accurate than the boosting type of classifier aggregation, as shown by Alfaro et al [8]. Furthermore, the Random Forest classifier [6] has been shown to have good accuracy compared to other classifiers [9] [10]. Therefore, we use boosting (specifically, Freund and Schapire's Adaboost algorithm [5]) and Random # Trace Forests for care product classification.…”
Section: Classifier Algorithmsmentioning
confidence: 99%
“…In turn, neural networks are less accurate than the boosting type of classifier aggregation, as shown by Alfaro et al [8]. Furthermore, the Random Forest classifier [6] has been shown to have good accuracy compared to other classifiers [9] [10]. Therefore, we use boosting (specifically, Freund and Schapire's Adaboost algorithm [5]) and Random # Trace Forests for care product classification.…”
Section: Classifier Algorithmsmentioning
confidence: 99%
“…These kinds of methods have gained a large acceptance in the machine learning community during the last two decades due to their high performance. Decision trees are the most common classifier structure considered and much work has been done in the topic [31,32], although they can be used with any other type of classifiers (the use of neural networks is also very extended, see for example [33]). …”
Section: Related Work On Mcssmentioning
confidence: 99%
“…The interested reader is referred to [32,33] for two reviews for the case of decision tree (both) and neural network ensembles (the latter), including exhaustive experimental studies.…”
Section: Related Work On Mcssmentioning
confidence: 99%
“…classifier, decision trees (DTs) are one of the most commonly used methods because they are efficient [3], [4]. Considerable work has been done to determine the effective ways for constructing diverse DTs so that the benefit of ensemble construction could be achieved.…”
mentioning
confidence: 99%
“…There are many ways, such as using different training sets and learning methods, one can adopt to construct diverse DTs. It is argued that DTs construction using different data is likely to maintain more diversity than other approaches [4]- [6] because function that a DT determines approximates from the training data. A number of methods have also been investigated to create different data sets for proper diversity.…”
mentioning
confidence: 99%