2019
DOI: 10.1109/access.2019.2923846
|View full text |Cite
|
Sign up to set email alerts
|

A Classification Method Based on Feature Selection for Imbalanced Data

Abstract: Imbalanced data are very common in the real world, and it may deteriorate the performance of the conventional classification algorithms. In order to resolve the imbalanced classification problems, we propose an ensemble classification method that combines evolutionary under-sampling and feature selection. We employ the Bootstrap method in original data to generate many sample subsets. V -statistic is developed to measure the distribution of imbalanced data, and it is also taken as the optimization objective of… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 54 publications
(25 citation statements)
references
References 45 publications
0
25
0
Order By: Relevance
“…For a fair evaluation, all of the methods examined in this section were selected from among wrapper-based methods. These wrapper-based methods include PSO-based [61], ACO-based [62], and ABC-based [63] 4…”
Section: Experiments and Discussionmentioning
confidence: 99%
“…For a fair evaluation, all of the methods examined in this section were selected from among wrapper-based methods. These wrapper-based methods include PSO-based [61], ACO-based [62], and ABC-based [63] 4…”
Section: Experiments and Discussionmentioning
confidence: 99%
“…The experiments have been performed in Waikato Environment for Knowledge Analysis (WEKA) [19]. WEKA, acknowledged as a landmark system in machine learning and DM, has become a widely used tool for DM research [20].…”
Section: Methodsmentioning
confidence: 99%
“…This feature selection technique works by measuring the uncertainty of a random variable x to another variable y as given by (6) and 7. P(xi) is the prior probability for all values of x and P(xi |yi) is the posterior probability of x given y [54], [55]. It is claimed to be effective in feature selection for large scale data sets.…”
Section: Symmetrical Uncertainty (Su)mentioning
confidence: 99%