2017
DOI: 10.15439/2017f536
|View full text |Cite
|
Sign up to set email alerts
|

A case study on machine learning model for code review expert system in software engineering

Abstract: Abstract-Code review is a key tool for quality assurance in software development. It is intended to find coding mistakes overlooked during development phase and lower risk of bugs in final product. In large and complex projects accurate code review is a challenging task. As code review depends on individual reviewer predisposition there is certain margin of source code changes that is not checked as it should. In this paper we propose machine learning approach for pointing project artifacts that are significan… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 10 publications
0
7
0
Order By: Relevance
“…Pradhan et al use machine learning to predict software defects in large systems [95]. Madera and Tomon apply machine learning to identify source code artefacts that are probably endangered by software defects [96]. Software defect prediction is highly influenced by the amount of availability of data for training the machine learning models, including neural networks, SVM, KNN, K-Means Clustering, Naive Bayes, decision trees, logistic and linear regression models, as well as their combinations with ensemble learning [97].…”
Section: Adaptive Methods In Quality Modellingmentioning
confidence: 99%
“…Pradhan et al use machine learning to predict software defects in large systems [95]. Madera and Tomon apply machine learning to identify source code artefacts that are probably endangered by software defects [96]. Software defect prediction is highly influenced by the amount of availability of data for training the machine learning models, including neural networks, SVM, KNN, K-Means Clustering, Naive Bayes, decision trees, logistic and linear regression models, as well as their combinations with ensemble learning [97].…”
Section: Adaptive Methods In Quality Modellingmentioning
confidence: 99%
“…Moreover, imbalanced sizes of software projects and datasets were also pointed out to be major obstacles in evaluating the techniques empirically [70,207]. Lack of generalizability and overfitting problems appeared to be the highest limitation in the articles as the ML models have shown fewer results when applied to diverse cross-project datasets [122,144]. Future directions include improvement of precision while maintaining recall in ML models [70].…”
Section: Discussionmentioning
confidence: 99%
“…As observed in some studies, e.g., [140,176], the lack of generalizability regarded as over-fitting problems is one of few major limiting factors, which decreases the accuracy of results. This leads to lesser results when ML models are applied to diverse cross-project datasets.…”
Section: B Limitationsmentioning
confidence: 99%