Research and Development in Intelligent Systems XVII 2001
DOI: 10.1007/978-1-4471-0269-4_5
|View full text |Cite
|
Sign up to set email alerts
|

Selecting Optimal Split-Functions for Large Datasets

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2004
2004
2022
2022

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 5 publications
0
2
0
Order By: Relevance
“…The training set was composed of 70% of all the input data and the test with the remaining 30%. Gain ration and Gini index were used to rank all features according to their correlation with each class [17,18]. Thus, we selected the five features that achieved the highest scores for classification within each machine learning classifier and each operator.…”
Section: Machine Learning Classificationmentioning
confidence: 99%
“…The training set was composed of 70% of all the input data and the test with the remaining 30%. Gain ration and Gini index were used to rank all features according to their correlation with each class [17,18]. Thus, we selected the five features that achieved the highest scores for classification within each machine learning classifier and each operator.…”
Section: Machine Learning Classificationmentioning
confidence: 99%
“…Two widely used split criterion is Gini Index and Information Gain. A lot of research was dedicated to understand which of them produce the best decision tree for a given dataset [46], [38], [70], [55]. Although most of empirical studies concluded that there is no significant differences between those two criteria and the disagreement is generally no higher than 2% of all cases, we show the validation accuracy achieved with different feature sets when using different criteria with Extra-Trees classifier (n estimators = 700) in Fig.…”
Section: H Parameter Tuning In Extra-trees Classifiermentioning
confidence: 99%