1997
DOI: 10.1142/s021821309700027x
|View full text |Cite
|
Sign up to set email alerts
|

Data Mining Using $\mathcal{MLC}++$ a Machine Learning Library in C++

Abstract: Data mining algorithms including maching learning, statistical analysis, and pattern recognition techniques can greatly improve our understanding of data warehouses that are now becoming more widespread. In this paper, we focus on classification algorithms and review the need for multiple classification algorithms. We describe a system called [Formula: see text], which was designed to help choose the appropriate classification algorithm for a given dataset by making it easy to compare the utility of different … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
70
0
19

Year Published

1998
1998
2022
2022

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 136 publications
(92 citation statements)
references
References 6 publications
0
70
0
19
Order By: Relevance
“…The landmarking meta-features extracted from the meta-data of these dataset. Also the Naive Bayes [17], IBK [18] , J48 [19], AdaBoost , LogitBoost [20], PART [21], RandomForest [22], Bagging [23] and SMO [24], classifier are apply on given dataset and accuracy of each classifier is calculated. This is knowledge base of system.…”
Section: System Outlinementioning
confidence: 99%
See 1 more Smart Citation
“…The landmarking meta-features extracted from the meta-data of these dataset. Also the Naive Bayes [17], IBK [18] , J48 [19], AdaBoost , LogitBoost [20], PART [21], RandomForest [22], Bagging [23] and SMO [24], classifier are apply on given dataset and accuracy of each classifier is calculated. This is knowledge base of system.…”
Section: System Outlinementioning
confidence: 99%
“…Where, the neighbor selection [17] algorithm is used with traditional distance formula. The distance of new dataset from the old dataset is calculated by, distance = ∑(new meta-feature)-∑(old meta-feature).…”
Section: Experiment-1mentioning
confidence: 99%
“…• Validación cruzada: se ha realizado una validación cruzada de k partes [14] donde k = 10. De esta manera, nuestro conjunto de datos se divide 10 veces en 10 partes diferentes.…”
Section: Metodología De Experimentaciónunclassified
“…For each domain, we induced classifiers for the minority class (for Road, we chose the class Grass). We selected several induction algorithms from MLC++ (Kohavi, Sommerfield, & Dougherty, 1997): a decision tree learner (MC4), Naive Bayes with discretization (NB), k-nearest neighbor for several k values (IBk), and Bagged-MC4 (Breiman, 1996). MC4 is similar to C4.5 (Quinlan, 1993); probabilistic predictions are made by using a Laplace correction at the leaves.…”
Section: Our Studymentioning
confidence: 99%