2017
DOI: 10.1007/978-3-319-57240-6_18
|View full text |Cite
|
Sign up to set email alerts
|

Cost-Complexity Pruning of Random Forests

Abstract: Random forests perform boostrap-aggregation by sampling the training samples with replacement. This enables the evaluation of out-of-bag error which serves as a internal crossvalidation mechanism. Our motivation lies in using the unsampled training samples to improve each decision tree in the ensemble. We study the effect of using the out-of-bag samples to improve the generalization error first of the decision trees and second the random forest by post-pruning. A prelimiary empirical study on four UCI reposito… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 13 publications
0
4
0
Order By: Relevance
“…Random Forest: Random Forest is an ensemble learning method of using bagging and random features selection to construct a multitude of decision trees during the training [38], [40]. This classification algorithm is widely used in data mining area.…”
Section: B Baseline Comparison Methodsmentioning
confidence: 99%
“…Random Forest: Random Forest is an ensemble learning method of using bagging and random features selection to construct a multitude of decision trees during the training [38], [40]. This classification algorithm is widely used in data mining area.…”
Section: B Baseline Comparison Methodsmentioning
confidence: 99%
“…This process is parametrized by the complexity parameter, α, that indicates a particular tree dimension. How α is calculated is beyond the scope of this work, but more information on the MCCP can be found at [6]. The more leaf nodes a tree has, the higher its complexity becomes and the lower the value of α.…”
Section: Fig 1: Schematic Of 2d Fe Model (Left -L) and Multi-spot Sen...mentioning
confidence: 99%
“…It should be noted that these methods are advantageous in a large set, but contraindicated in a moderate size set. Thus, different optimization techniques have been employed and various techniques are utilized to overcome this situation including genetic algorithm (Ko et al, 2014;Mousavi et al, 2018;Taleb Zouggar and Adla, 2019), Particle Swarm (Escalante et al, 2010(Escalante et al, , 2012Taghavi et al, 2015), greedy algorithm (Guo and Fan, 2011;Dai, 2013;Dai and Li, 2015), semi-definite programming (Zhang et al, 2006), quadratic programming (Li and Zhou, 2009), hill climbing (Partalas et al, 2010;Indyk et al, 2014), localized generalization error (Pratama et al, 2018), bi-objective evolutionary optimization (Yin et al, 2014;Qian et al, 2015) and lately the cost-complexity pruning (Kiran and Serra, 2017;Wang et al, 2017;Fernandes et al, 2017).…”
Section: The Optimization-based Pruningmentioning
confidence: 99%