2016
DOI: 10.9781/ijimai.2016.419
|View full text |Cite
|
Sign up to set email alerts
|

Statistical Comparisons of the Top 10 Algorithms in Data Mining for Classi cation Task

Abstract: -This work is builds on the study of the 10 top data mining algorithms identified by the IEEE International Conference on Data Mining (ICDM) community in December 2006. We address the same study, but with the application of statistical tests to establish, a more appropriate and justified ranking classifier for classification tasks. Current studies and practices on theoretical and empirical comparison of several methods, approaches, advocated tests that are more appropriate. Thereby, recent studies recommend a … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
33
0
1

Year Published

2016
2016
2023
2023

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 61 publications
(34 citation statements)
references
References 7 publications
0
33
0
1
Order By: Relevance
“…Presented results support the neural networks capacity to solve statistical [39] and parameter estimation [40] problems. The current proposal applied a novel feature extraction method by using histograms of groups of K samples, which was combined with deep learning and classifier fusion approaches to produce an improved K parameter estimator.…”
Section: B Significance Of Resultsmentioning
confidence: 73%
“…Presented results support the neural networks capacity to solve statistical [39] and parameter estimation [40] problems. The current proposal applied a novel feature extraction method by using histograms of groups of K samples, which was combined with deep learning and classifier fusion approaches to produce an improved K parameter estimator.…”
Section: B Significance Of Resultsmentioning
confidence: 73%
“…Our method, however, uses many different information sources about the individual cases without assuming any relationship between these factors and transmission. Additionally, our method uses naive Bayes, a simple but powerful machine learning tool that has many diverse applications (33,(49)(50)(51)(52). Although traditionally a naive Bayes model is trained with a set of true events, our method performs almost as well when SNP distance is used as a transmission proxy.…”
Section: Discussionmentioning
confidence: 99%
“…K-means is one of the well-known algorithms for clustering [28], [29] of which various modifications have been proposed including fuzzy logic [30]. In this algorithm each cluster is characterized by its center point i.e.…”
Section: A Integrated Scheme Of Customer Segmentation With Campaign mentioning
confidence: 99%