Software effort estimation accuracy is a key factor in effective planning, controlling, and delivering a successful software project within budget and schedule.The overestimation and underestimation both are the key challenges for future software development, henceforth there is a continuous need for accuracy in software effort estimation. The researchers and practitioners are striving to identify which machine learning estimation technique gives more accurate results based on evaluation measures, datasets and other relevant attributes. The authors of related research are generally not aware of previously published results of machine learning effort estimation techniques. The main aim of this study is to assist the researchers to know which machine learning technique yields the promising effort estimation accuracy prediction in software development. In this article, the performance of the machine learning ensemble and solo techniques are investigated on publicly and non-publicly domain datasets based on the two most commonly used accuracy evaluation metrics. We used the systematic literature review methodology proposed by Kitchenham and Charters. This includes searching for the most relevant papers, applying quality assessment (QA) criteria, extracting data, and drawing results. We have evaluated a state-of-the-art accuracy performance of 35 selected studies (17 ensemble, 18 solo) using mean magnitude of relative error and PRED (25) as a set of reliable accuracy metrics for performance evaluation of accuracy among two techniques to report the research questions stated in this study. We found that machine learning techniques are the most frequently implemented in the construction of ensemble effort estimation (EEE) techniques. The results of this study revealed that the EEE techniques usually yield a promising estimation accuracy than the solo techniques.
In recent years, due to significant evolution in adopting new technologies and development methodologies in the field of software engineering, the developers and researchers are striving to optimize the accuracy of software effort estimation (SEE). The overestimation and underestimation both are the key challenges for software progress; hence, there is a continuous need for an accurate SEE. This paper highlights a systematic review of studies associated with the best practices of use case point (UCP) and expert judgment–based software development effort estimation techniques. The primary aim and contribution of this paper are to support the researchers through an extensive review to ease to other researcher's search for effort estimation studies. We have performed state‐of‐the‐art review from five viewpoints of reference: (a) review of studies concerning UCPs and expert judgment–based effort estimation, (b) research contribution and future recommendation in different novelties, (c) usage of the dataset, (d) availability of accuracy metrics, and (e) findings of the studies. We have performed a systematic review of studies which are published in the period of 2000 to 2019. We have selected a total of 34 primary studies of UCP and expert judgment–based estimation techniques to report the research questions stated in this review.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.