2022
DOI: 10.1002/cpe.7586
|View full text |Cite
|
Sign up to set email alerts
|

Undersampling of approaching the classification boundary for imbalance problem

Abstract: Using imbalanced data in classification affect the accuracy. If the classification is based on imbalanced data directly, the results will have large deviations. A common approach to dealing with imbalanced data is to re-structure the raw dataset via undersampling method. The undersampling method usually uses random or clustering approaches to trimming the majority class in the dataset, since some data in the majority class makes not contribute to classification model. In this paper a revised undersampling appr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 41 publications
0
4
0
Order By: Relevance
“…A healthy number of intrepid researchers have applied oversampling [9][10][11][12][13][14], undersampling [15][16][17][18][19], and hybrid [20][21][22][23] preprocessing methods to restore balance to their training datasets. These methods are combined with feature classification methods to maximize benefits.…”
Section: A Data-level Mitigation Effortsmentioning
confidence: 99%
“…A healthy number of intrepid researchers have applied oversampling [9][10][11][12][13][14], undersampling [15][16][17][18][19], and hybrid [20][21][22][23] preprocessing methods to restore balance to their training datasets. These methods are combined with feature classification methods to maximize benefits.…”
Section: A Data-level Mitigation Effortsmentioning
confidence: 99%
“…The ensemble of the α− Trees framework (EAT) uses underbagging technology to achieve good results when applied to imbalanced classification problems [26]. In terms of boosting, the technologies commonly used include RUSBoost [27], SMOTEBoost [28], CSBBoost [29], Adaboost, and AsymBoost [30]. In addition, Yang et al [31] employed progressive density-based weighted ensemble learning, and Ren et al [32] designed a weighted integration scheme that was obtained via the use of classifiers based on the original imbalanced data set for ensemble learning.…”
Section: Ensemble Learning In Undersamplingmentioning
confidence: 99%
“…Combining these two methods can result in more accurate data for model building, which can improve the outcomes. The problem of data imbalance has been overlooked in petroleum engineering research, but it can cause results to favor the majority class [61]. However, the minority classes often require more accuracy.…”
Section: Potential Research Opportunitiesmentioning
confidence: 99%