2021
DOI: 10.1016/j.mlwa.2021.100170
|View full text |Cite
|
Sign up to set email alerts
|

Assessing feature selection method performance with class imbalance data

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
25
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(25 citation statements)
references
References 16 publications
0
25
0
Order By: Relevance
“…A simulation study was conducted to examine the effect of informative and non-informative features on feature selection techniques. Like [4], synthetic data of 30 features with varying imbalance rates and noise (i.e. non-informative features), were generated using the "make classification" library in python's scikit-learn.datasets.…”
Section: Methods and Experimental Designmentioning
confidence: 99%
See 3 more Smart Citations
“…A simulation study was conducted to examine the effect of informative and non-informative features on feature selection techniques. Like [4], synthetic data of 30 features with varying imbalance rates and noise (i.e. non-informative features), were generated using the "make classification" library in python's scikit-learn.datasets.…”
Section: Methods and Experimental Designmentioning
confidence: 99%
“…filter, wrapper, and embedded methods, are widely been used. The filters measure the relevance (e.g., Fisher score, mutual information, correlation [3][4][5]) of a feature with the target. Filters are computationally fast, simple, easy to scale, and generally don't require a learning model [6].…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…WS/WP is usually predicted based on its historical data and the meteorological data which are commonly represented by extensive time series with multi-dimensions, data noises, and redundancy or lack of attributes. Moreover, using a very complex AI-based forecasting model with a huge volume of input data to train the model is not always feasible because of the limited calculation time and memory [5], [6]. Therefore, approaches for reducing the number of features in the model to eliminate extraneous and redundant data, avoid the loss of relevant information, and reduce computation time are necessary.…”
Section: Introductionmentioning
confidence: 99%