2020
DOI: 10.1016/j.eswa.2019.113026
|View full text |Cite
|
Sign up to set email alerts
|

Understanding the apparent superiority of over-sampling through an analysis of local information for class-imbalanced data

Abstract: Data plays a key role in the design of expert and intelligent systems and therefore, data preprocessing appears to be a critical step to produce high-quality data and build accurate machine learning models. Over the past decades, increasing attention has been paid towards the issue of class imbalance and this is now a research hotspot in a variety of fields. Although the resampling methods, either by undersampling the majority class or by over-sampling the minority class, stand among the most powerful techniqu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

2
26
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 72 publications
(28 citation statements)
references
References 94 publications
2
26
0
Order By: Relevance
“…Over and under sampling strategies are very popular and effective approaches to deal with the class imbalance problem [21,25,33,50]. To compensate the class imbalance by biasing the process of discrimination, the ROS algorithm randomly replicates samples from the minority classes while the RUS technique randomly eliminates samples from the majority classes, until achieving a relative classes balance [23,60].…”
Section: Sampling Class Imbalance Approachesmentioning
confidence: 99%
See 4 more Smart Citations
“…Over and under sampling strategies are very popular and effective approaches to deal with the class imbalance problem [21,25,33,50]. To compensate the class imbalance by biasing the process of discrimination, the ROS algorithm randomly replicates samples from the minority classes while the RUS technique randomly eliminates samples from the majority classes, until achieving a relative classes balance [23,60].…”
Section: Sampling Class Imbalance Approachesmentioning
confidence: 99%
“…Hybrid methods generally employ SMOTE to compensate the class imbalance, because this method reduces the possibilities of over-training or over-fitting [1]. They use methods based in nearest neighbor rule to reduce overlap or noise in the dataset [25].…”
Section: Hybrid Sampling Class Imbalance Strategiesmentioning
confidence: 99%
See 3 more Smart Citations