2008 the Third International Multi-Conference on Computing in the Global Information Technology (Iccgi 2008) 2008
DOI: 10.1109/iccgi.2008.34
|View full text |Cite
|
Sign up to set email alerts
|

Dimensionality Reduction for Feature and Pattern Selection in Classification Problems

Abstract: Reducing the dimensionality of a dataset is an important and often challenging task. This can be done by either reducing the number of features, a task called feature selection, or by reducing the number of patterns, called data reduction. In this paper we propose methods that employ a novel concept called Discernibility for achieving these two tasks separately, with the aim to solve classification problems. The experimental results verify our claim that the proposed methods are a viable alternative for dimens… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 25 publications
(36 reference statements)
0
2
0
Order By: Relevance
“…Second, descriptors were reduced using the Boruta method (Kursa et al, 2010). We used four different machine learning methods, namely, Support Vector Machine (SVM) (Mitchell, 1997), random forest (Breiman, 2001), Extreme Gradient Boosting (XGBoost) (Chen and Guestrin, 2016), and kappa nearest neighbor (kNN) (Voulgaris and Magoulas, 2008), to build the classification models. All the classification experiments and calculations were conducted using the R.3.0.2 environment (http://www.R-project.org/) and Python (http://www.python.org/) platform.…”
Section: Data Collection and Model Buildingmentioning
confidence: 99%
“…Second, descriptors were reduced using the Boruta method (Kursa et al, 2010). We used four different machine learning methods, namely, Support Vector Machine (SVM) (Mitchell, 1997), random forest (Breiman, 2001), Extreme Gradient Boosting (XGBoost) (Chen and Guestrin, 2016), and kappa nearest neighbor (kNN) (Voulgaris and Magoulas, 2008), to build the classification models. All the classification experiments and calculations were conducted using the R.3.0.2 environment (http://www.R-project.org/) and Python (http://www.python.org/) platform.…”
Section: Data Collection and Model Buildingmentioning
confidence: 99%
“…DR can be done via either reduce the number of features, a task called feature selection, or via reduce the number of patterns, called data reduction [19]. Data dimensionality reduction produce a compact low-dimensional encoding of a given high dimensional data set [20].…”
Section: Dimension Reduction (Dr) and Feature Selectionmentioning
confidence: 99%