2002
DOI: 10.1109/34.990132
|View full text |Cite
|
Sign up to set email alerts
|

Complexity measures of supervised classification problems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
53
0

Year Published

2009
2009
2020
2020

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 511 publications
(56 citation statements)
references
References 20 publications
3
53
0
Order By: Relevance
“…However, in order to describe the real complexity of the problem and since our eventual goal is to study the behavior of various data mining methods for prediction, we need to find other measures that are independent of such choices. Previous works highlight the idea that a single descriptor may not be sufficient (Tin Kam & Basu, 2002).…”
Section: Datasets Complexitymentioning
confidence: 99%
“…However, in order to describe the real complexity of the problem and since our eventual goal is to study the behavior of various data mining methods for prediction, we need to find other measures that are independent of such choices. Previous works highlight the idea that a single descriptor may not be sufficient (Tin Kam & Basu, 2002).…”
Section: Datasets Complexitymentioning
confidence: 99%
“…According to Ho and Basu [11], the difficulty of a classification problem can be attributed to three main aspects: the ambiguity among the classes, the complexity of the separation between the classes, and the data sparsity and dimensionality. Usually, there is a combination of these aspects.…”
Section: Complexity Indices For Describing Datamentioning
confidence: 99%
“…For such, we employ a series of statistic and geometric descriptors described in [11]. These indices account for the difficulty of a classification problem by analyzing some characteristics of the dataset and the predictive performance of some simple classification models induced using the dataset.…”
Section: Introductionmentioning
confidence: 99%
“…On the other hand, assessing the behavior of the dataset ECU, establishing the most relevant attributes, and reducing any kind of data noise or inconsistency are vital actions [21]. For this, some filters are used, as explained in [22, 23, 24, and 25], allowing more trustworthiness in subsequent intrusion detection stages.…”
Section: ) Data Normalizationmentioning
confidence: 99%