2018 44th Euromicro Conference on Software Engineering and Advanced Applications (SEAA) 2018
DOI: 10.1109/seaa.2018.00048
|View full text |Cite
|
Sign up to set email alerts
|

An Exploratory Study of Search Based Training Data Selection for Cross Project Defect Prediction

Abstract: Context: Search based approaches are gaining attention in cross project defect prediction (CPDP). The complexity of such approaches and existence of various design decisions are important issues to consider. Objective: We aim at investigating factors that can affect the performance of search based selection (SBS) approaches. We study a genetic instance selection approach (GIS) and present an evaluation of design options for search based CPDP. Method: Using an exploratory approach, data from different options o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 26 publications
(38 reference statements)
0
4
0
Order By: Relevance
“…• recall [86] • accuracy [129] • probability of false positive [65] • true negative rate [121] • balance [95] • ROC-AUC [76] • F-Measure [71] • G-Measure [56] • H-Measure [57] • G-Mean [82] • Win/Draw/Loss [59] • false negative rate [121] • Matthews correlation coefficient [121] FIGURE 13. Evaluation measures frequency Figure 13 shows that precision, recall, ROC-AUC, and F-measure are the preferred evaluation measures.…”
Section: E Rq5: Which Are the Evaluation Measures Applied To Cpdp Mod...mentioning
confidence: 99%
“…• recall [86] • accuracy [129] • probability of false positive [65] • true negative rate [121] • balance [95] • ROC-AUC [76] • F-Measure [71] • G-Measure [56] • H-Measure [57] • G-Mean [82] • Win/Draw/Loss [59] • false negative rate [121] • Matthews correlation coefficient [121] FIGURE 13. Evaluation measures frequency Figure 13 shows that precision, recall, ROC-AUC, and F-measure are the preferred evaluation measures.…”
Section: E Rq5: Which Are the Evaluation Measures Applied To Cpdp Mod...mentioning
confidence: 99%
“…F-Measure is selected in this study as the basis of our selection of the best approach. Additionally, a combination of F-measure and GMean, i.e., F×GMean is used further as the basis for fitness assignment for LSH parameter tuning, learning technique hyper-parameter tuning, [5,6,25].…”
Section: Performance Measures and Toolsmentioning
confidence: 99%
“…Researchers build their prediction models based on software metrics derived from source code repository (e.g., Change metrics [7], CK metrics [20], Object-oriented metrics [9]) using machine learning classifiers (e.g., Naive Bayes [21], Support Vector Machine [22], Decision Tree [23], Random Forest [24]) to classify faulty and non-faulty modules. The main challenge of CPDP is to reduce data divergence between source and target projects data sets.…”
Section: A Statistical and Machine Learning Based Cpdpmentioning
confidence: 99%