2022
DOI: 10.1007/s40745-021-00366-5
|View full text |Cite
|
Sign up to set email alerts
|

Feature Selection in Imbalanced Data

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(9 citation statements)
references
References 22 publications
0
9
0
Order By: Relevance
“…Feature selection in imbalanced data 50 Kamalov and Thabtah explored feature selection in imbalanced data. However, their work primarily addresses class imbalance rather than the broader challenges of high-dimensional feature selection.…”
Section: In-depth Review Of Existing Machine Learning Models Used For...mentioning
confidence: 99%
“…Feature selection in imbalanced data 50 Kamalov and Thabtah explored feature selection in imbalanced data. However, their work primarily addresses class imbalance rather than the broader challenges of high-dimensional feature selection.…”
Section: In-depth Review Of Existing Machine Learning Models Used For...mentioning
confidence: 99%
“…For each of the 600 features, we apply a decision tree (DT) classifier to the training portion of data frame D. This classifier is then tested on the corresponding test portion of D to obtain an F1-score (lines 12–17 in Algorithm A1). The F1-score represents how well a feature performs in categorizing the positive (minority) data, making it an important indicator when the underlying data distribution is unbalanced [ 48 ]. After completing total, I iterations, the average F1-score is computed for all 600 features to establish their rankings (lines 19–21 in Algorithm A1).…”
Section: Appendix A1 Model Building Algorithmmentioning
confidence: 99%
“…Finally, we have identified and reported the most relevant set of magnetic field parameters for the prediction of major/minor flare classes from multiple partitions of the SWAN-SF data set. Selecting the most relevant and informative parameters can help reduce prediction errors by avoiding overfitting (Kamalov et al 2022). Feature selection is also an effective measure for data compression and visualization (Filali Boubrahimi & Hamdi 2022).…”
Section: Introductionmentioning
confidence: 99%