2019
DOI: 10.1007/978-3-030-19738-4_36
|View full text |Cite
|
Sign up to set email alerts
|

Multi Sampling Random Subspace Ensemble for Imbalanced Data Stream Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 24 publications
0
4
0
Order By: Relevance
“…(1) Predicting death from hospital-acquired infections in trauma patients in the absence of a balanced dataset (C5.0 and CHAID); (2) Predicting death from hospital-acquired infection in the trauma patients using a balanced dataset by sampling methods (reduced data set) (C5.0 and CHAID); (3) Clustering hospital-acquired infections in trauma patients by k-means algorithms; (4) Predicting death from hospital-acquired infections in trauma patients in each cluster (C5.0 and CHAID); (5) Predicting death from hospital-acquired infections in trauma patients with SMOTE-C5.0 and ADASYN-C5.0; (6) Predicting death from hospital-acquired infections in the trauma patients with SMOTE-SVM, ADASYN-SVM, SMOTE-ANN, and ADASYN-ANN. Many previous studies have attempted to handle unbalanced data [12][13][14] by adopting various approaches, such as using the right evaluation metrics, resampling the training set (under-sampling, and over-sampling), using K-fold cross-validation appropriately, ensemble different resampled datasets, resampling different ratios, and clustering the frequent class. However, no best model for these problems has been identified, while this strongly relates to techniques, models, and subjects used [2].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…(1) Predicting death from hospital-acquired infections in trauma patients in the absence of a balanced dataset (C5.0 and CHAID); (2) Predicting death from hospital-acquired infection in the trauma patients using a balanced dataset by sampling methods (reduced data set) (C5.0 and CHAID); (3) Clustering hospital-acquired infections in trauma patients by k-means algorithms; (4) Predicting death from hospital-acquired infections in trauma patients in each cluster (C5.0 and CHAID); (5) Predicting death from hospital-acquired infections in trauma patients with SMOTE-C5.0 and ADASYN-C5.0; (6) Predicting death from hospital-acquired infections in the trauma patients with SMOTE-SVM, ADASYN-SVM, SMOTE-ANN, and ADASYN-ANN. Many previous studies have attempted to handle unbalanced data [12][13][14] by adopting various approaches, such as using the right evaluation metrics, resampling the training set (under-sampling, and over-sampling), using K-fold cross-validation appropriately, ensemble different resampled datasets, resampling different ratios, and clustering the frequent class. However, no best model for these problems has been identified, while this strongly relates to techniques, models, and subjects used [2].…”
Section: Introductionmentioning
confidence: 99%
“…Many previous studies have attempted to handle unbalanced data [ 12 14 ] by adopting various approaches, such as using the right evaluation metrics, resampling the training set (under-sampling, and over-sampling), using K-fold cross-validation appropriately, ensemble different resampled datasets, resampling different ratios, and clustering the frequent class. However, no best model for these problems has been identified, while this strongly relates to techniques, models, and subjects used [ 2 ].…”
Section: Introductionmentioning
confidence: 99%
“…The previous components are regularly updated to make the ensemble react to different kinds of concept drift. Klikowski et al 129 proposed a multi‐sampling random subspace ensemble method (MSRS). The algorithm uses a random subspace method and uses various sampling methods to balance the data to ensure proper diversity of classifier ensemble, and through the use of various oversampling techniques to ensure diversity.…”
Section: Passive Handling Methodsmentioning
confidence: 99%
“…The most popular approach lies in combining resampling techniques with Online Bagging (Wang et al, 2015Wang and Pineau, 2016). Similar strategies can be applied to Adaptive Random Forest (Gomes et al, 2017), Online Boosting (Klikowski and Woźniak, 2019;Gomes et al, 2019) 2017), Dynamic Feature Selection (Wu et al, 2014), Adaptive Random Forest with resampling (Ferreira et al, 2019), Kappa Updated Ensemble (Cano and Krawczyk, 2020), Robust Online Self-Adjusting Ensemble (Cano and Krawczyk, 2022) or any ensemble that can incrementally update its base learners (Ancy and Paulraj, 2020;Li et al, 2020). It is interesting to note that preprocessing approaches enhance diversity among base classifiers (Zyblewski et al, 2019).…”
Section: Ensembles For Imbalanced Data Streamsmentioning
confidence: 99%