2006
DOI: 10.1007/11731139_15
|View full text |Cite
|
Sign up to set email alerts
|

Boosting Prediction Accuracy on Imbalanced Datasets with SVM Ensembles

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
57
0

Year Published

2006
2006
2021
2021

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 114 publications
(58 citation statements)
references
References 11 publications
1
57
0
Order By: Relevance
“…SVM is a widely used machine learning method which has been applied to many real world problems providing satisfactory results. SVM works effectively with balanced dataset but provides suboptimal classification models considering the imbalanced dataset; several examples demonstrate this conclusion (Veropoulus et al, 1999;Akbani et al, 2004;Wu & Chang, 2003;Wu & Chang, 2005;Raskutti & Kowalczyk, 2004;Imam et al, 2006;Zou et al, 2008;Lin et al, 2009;Kang & Cho, 2006;Liu et al, 2006;Haibo & Garcia, 2009). SVM is biased toward the majority class and provides poor results concerning the minority class.…”
Section: Fuzzy Based Approachesmentioning
confidence: 98%
“…SVM is a widely used machine learning method which has been applied to many real world problems providing satisfactory results. SVM works effectively with balanced dataset but provides suboptimal classification models considering the imbalanced dataset; several examples demonstrate this conclusion (Veropoulus et al, 1999;Akbani et al, 2004;Wu & Chang, 2003;Wu & Chang, 2005;Raskutti & Kowalczyk, 2004;Imam et al, 2006;Zou et al, 2008;Lin et al, 2009;Kang & Cho, 2006;Liu et al, 2006;Haibo & Garcia, 2009). SVM is biased toward the majority class and provides poor results concerning the minority class.…”
Section: Fuzzy Based Approachesmentioning
confidence: 98%
“…Batista et al [3] proposed to apply SMOTE after performing a data cleaning (i.e., under-sampling) method such as Tomek links and the Wilson's Edited Nearest Neighbor Rule. Liu at al [14] over-sampled the minority class with SMOTE to some extent, then under-sampled the majority class a number of times to create bootstrap samples having the same or similar size with the over-sampled minority class.…”
Section: Related Workmentioning
confidence: 99%
“…We trained two classifiers using the R language for statistical computing (Liu et al 2006): an ensemble of Support Vector Machines (EnsSVM), and a weighted random forest (WRF). Two additional classifiers, namely, a (regular) SVM with an RBF kernel and a decision tree with a weighted loss function, were trained on the route blockage data set, but EnsSVM and WRF outperformed them in terms of maximizing recall.…”
Section: Ensemble Classifiersmentioning
confidence: 99%