2019
DOI: 10.2991/ijcis.d.191114.002
|View full text |Cite
|
Sign up to set email alerts
|

A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTE

Abstract: Imbalance learning is a challenging task for most standard machine learning algorithms. The Synthetic Minority Oversampling Technique (SMOTE) is a well-known preprocessing approach for handling imbalanced datasets, where the minority class is oversampled by producing synthetic examples in feature vector rather than data space. However, many recent works have shown that the imbalanced ratio in itself is not a problem and deterioration of the model performance is caused by other reasons linked to the minority cl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 47 publications
(25 citation statements)
references
References 46 publications
0
24
0
1
Order By: Relevance
“…Usually in order not to avoid elimination of significant majority of instances, Oversampling algorithms are preferred, and Synthetic Minority Oversampling Technique (SMOTE) algorithm proposed by Chawla et al ( 2002 ) is the most widely used. Subsequently more than 85 variants of SMOTE have been reported in literature to further improve the basic form of SMOTE in terms of different classification metrics (Fernández et al 2018 ) like borderline-SMOTE1 and borderline-SMOTE2, advanced SMOTE (A-SMOTE), Distributed version of SMOTE (Han et al 2005 ; Hooda and Mann 2019 ; Hussein 2019 ), etc. There seems to be only few literature reports dealing with detailed critical comparison of these proposed methods (Bajer et al 2019 ; Kovács 2019 ).…”
Section: Background Literaturementioning
confidence: 99%
“…Usually in order not to avoid elimination of significant majority of instances, Oversampling algorithms are preferred, and Synthetic Minority Oversampling Technique (SMOTE) algorithm proposed by Chawla et al ( 2002 ) is the most widely used. Subsequently more than 85 variants of SMOTE have been reported in literature to further improve the basic form of SMOTE in terms of different classification metrics (Fernández et al 2018 ) like borderline-SMOTE1 and borderline-SMOTE2, advanced SMOTE (A-SMOTE), Distributed version of SMOTE (Han et al 2005 ; Hooda and Mann 2019 ; Hussein 2019 ), etc. There seems to be only few literature reports dealing with detailed critical comparison of these proposed methods (Bajer et al 2019 ; Kovács 2019 ).…”
Section: Background Literaturementioning
confidence: 99%
“…La suma de estos valores es de 40 y se divide por el número de filas de la tabla, para este caso es 18. El resultado es 2.222 y al ser el promedio más bajo, ocupa el número 1 en el ranking [30].…”
Section: Resultados Y Discusiónunclassified
“…Synthetic instances that are far from the borderline are easier to categorize than those that are near to the borderline, which present a significant learning difficulty for the majority of classifiers. The authors in [ 32 ] describe an advanced strategy (A-SMOTE) for preprocessing imbalanced training sets based on these findings. It aims to clearly characterize the borderline and create pure synthetic samples from SMOTE generalization.…”
Section: Proposed Methodologymentioning
confidence: 99%
“…AdaBoost makes it possible to merge various “weak classifiers” into a single classifier which is called “strong classifier.” Decision trees with one level, or decision trees with only one split, are the most popular algorithm used with AdaBoost. Decision Stump is another name for these trees [ 32 ]. This approach creates a model by assigning equal weights to all of the data points.…”
Section: Exploratory Knowledgementioning
confidence: 99%