An essential goal for programmers is to minimize the cost of identifying and correcting defects in source code. Code review is commonly used for identifying programming defects. However, manual code review has some shortcomings:(1) it is time-consuming and (2) outcomes are subjective and depend on the skills of reviewers. An automated approach for assisting in code reviews is thus highly desirable. We present a tool for assisting in code review and results from our experiments evaluating the tool in different scenarios. The tool leveraged content available from professional programmer support forums (eg, StackOverflow.com) to determine potential defectiveness of a given piece of source code.The defectiveness is expressed on the scale of {Likely defective, neutral, unlikely to be defective}. The basic idea employed in the tool is (1) to identify a set P of discussion posts on Stack Overflow such that each p ∈ P contains source code fragment(s), which sufficiently resemble the input code C being reviewed, and (2) to determine the likelihood of C being defective by considering all p ∈ P.A novel aspect of our approach is to use document fingerprinting for comparing two pieces of source code. Our choice of document fingerprinting technique is inspired by source code plagiarism detection tools where it has proven to be very successful. In the experiments that we performed to verify the effectiveness of our approach, source code samples from more than 300 GitHub open-source repositories were taken as input. An F1 score of 0.94 has been achieved in identifying correct/relevant results. KEYWORDS automated software engineering, code review, crowd knowledge, software development, Stack Overflow
INTRODUCTIONWhat are we trying to do and why is it important?We present a novel tool that assists in carrying out effective code reviews. Identifying and fixing buggy code consume significant time and resources in a software development project. Code review by peers 1 and experienced programmers is an effective method 2,3 for identifying potentially buggy codes. However, the process of code review is slow, and quality of results depends on the skills and experience of the reviewers involved. Moreover, a code review carried out by an Shipra Sharma and Balwinder Sodhi contributed equally to this work.