2019
DOI: 10.11591/ijece.v9i4.pp3241-3246
|View full text |Cite
|
Sign up to set email alerts
|

Ensemble learning for software fault prediction problem with imbalanced data

Abstract: Fault prediction problem has a crucial role in the software development process because it contributes to reducing defects and assisting the testing process towards fault-free software components. <span lang="EN-US">Therefore, there are a lot of efforts aiming to address this type of issues, in which static code characteristics are usually adopted to construct fault classification models. </span> One of the challenging problems influencing the performance of predictive classifiers is the high imbal… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 21 publications
(9 citation statements)
references
References 24 publications
0
9
0
Order By: Relevance
“…Using the undersampling method called "near miss," the dataset's distribution was modified. Data sampling methods aim to address class imbalance by manipulating the dataset, typically by removing the majority of class samples, to achieve a more balanced ISSN: 2502-4752  distribution [28], [29]. Among these methods, near miss is a technique used to tackle class imbalance by removing instances from the majority class.…”
Section: Class Imbalance and Data Sampling Methodsmentioning
confidence: 99%
“…Using the undersampling method called "near miss," the dataset's distribution was modified. Data sampling methods aim to address class imbalance by manipulating the dataset, typically by removing the majority of class samples, to achieve a more balanced ISSN: 2502-4752  distribution [28], [29]. Among these methods, near miss is a technique used to tackle class imbalance by removing instances from the majority class.…”
Section: Class Imbalance and Data Sampling Methodsmentioning
confidence: 99%
“…Khuat and Le [11] applied the SMOTE on a diabetes mellitus dataset for balancing the class distribution of the diabetes mellitus positive and negative class. According to Seo and Kim [12], the synthetic oversampling technique is one of the most powerful techniques widely employed in medicine for imbalanced class distribution.…”
Section: Related Workmentioning
confidence: 99%
“…In these studies, it is seen that NB (Arar & Ayan, 2017), Bayesian Network (BN; (Pandey et al, 2018), Support Vector Regression (SVR) (Kaur et al, 2017; Singh & Chaturvedi, 2013), Decision Tree (DT; Hammouri et al, 2018) and RF (Immaculate et al, 2019) algorithms have been preferred. Some ensemble learning studies used simple majority voting (Khuat & Le, 2019); Kumar et al (2017) and weighted majority voting (Moustafa et al, 2018) techniques for predicting software bugs. This article prefers NB, MLP, SVM, DT (C4.5), RF and AdaBoost algorithms as base learners for the proposed MVOC approach.…”
Section: Related Workmentioning
confidence: 99%