2021
DOI: 10.1016/j.infsof.2021.106664
|View full text |Cite
|
Sign up to set email alerts
|

The impact of using biased performance metrics on software defect prediction research

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
32
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 38 publications
(33 citation statements)
references
References 62 publications
1
32
0
Order By: Relevance
“…The experimental results show that MKEL method is better against total six methods under F‐1, including coding‐based ensemble learning (CEL) (Sun et al, 2012), dynamic AdaBoost.NC (S. Wang & Yao, 2013), weighted Naïve Bayes (T. Wang & Li, 2010), Compressed C4.5 decision tree (Wang et al, 2012), Cost‐sensitive Boosting Neural Network (CBNN) (Zheng, 2010), and Asymmetric Kernel Principal Component Classification (AKPCC) (Y. Ma et al, 2012). As F‐1 is problematic in defect prediction (Yao & Shepperd, 2021), the experimental results are sceptical. Most of these classical methods are associated with Boosting, but the evaluation indicators of these methods are not unified, which cannot provide a meaningful reference for industry personnel.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…The experimental results show that MKEL method is better against total six methods under F‐1, including coding‐based ensemble learning (CEL) (Sun et al, 2012), dynamic AdaBoost.NC (S. Wang & Yao, 2013), weighted Naïve Bayes (T. Wang & Li, 2010), Compressed C4.5 decision tree (Wang et al, 2012), Cost‐sensitive Boosting Neural Network (CBNN) (Zheng, 2010), and Asymmetric Kernel Principal Component Classification (AKPCC) (Y. Ma et al, 2012). As F‐1 is problematic in defect prediction (Yao & Shepperd, 2021), the experimental results are sceptical. Most of these classical methods are associated with Boosting, but the evaluation indicators of these methods are not unified, which cannot provide a meaningful reference for industry personnel.…”
Section: Related Workmentioning
confidence: 99%
“…Firstly, the class imbalance can reduce the accuracy and reliability of the defect model in CVDP and IVDP, as shown in the result analysis. And, class imbalances can distort some performance indicators about machine learning, for example, accuracy, precision, and F‐measure (Yao & Shepperd, 2021). Based on practical experience, we employ two basic and three comprehensive indicators to evaluate the performance of defect models, improving the reliability of the evaluation results in CVDP and IVDP.…”
Section: Threats To Validitymentioning
confidence: 99%
See 2 more Smart Citations
“…The unsupervised ML technique is applied when working on unlabelled datasets, unlike supervised ML [14]. However, the performance of an ML technique depends largely on the quality of the datasets used for training such an ML technique [47][48][49].…”
Section: Related Workmentioning
confidence: 99%