2022 IEEE International Conference on Software Testing, Verification and Validation Workshops (ICSTW) 2022
DOI: 10.1109/icstw55395.2022.00039
|View full text |Cite
|
Sign up to set email alerts
|

µBert: Mutation Testing using Pre-Trained Language Models

Abstract: We introduce µBERT, a mutation testing tool that uses a pre-trained language model (CodeBERT) to generate mutants. This is done by masking a token from the expression given as input and using CodeBERT to predict it. Thus, the mutants are generated by replacing the masked tokens with the predicted ones. We evaluate µBERT on 40 real faults from Defects4J and show that it can detect 27 out of the 40 faults, while the baseline (PiTest) detects 26 of them. We also show that µBERT can be 2 times more cost-effective … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(16 citation statements)
references
References 29 publications
0
16
0
Order By: Relevance
“…Table 1 summarises the type of targeted AST nodes by µBERT, with corresponding example expressions and induced mutants. We refer to these as the conventional mutations provided by µBERT, denoted by µBERT conv in our evaluation, previously introduced in the preliminary version of the approach [18].…”
Section: Ast Nodes Selectionmentioning
confidence: 99%
See 3 more Smart Citations
“…Table 1 summarises the type of targeted AST nodes by µBERT, with corresponding example expressions and induced mutants. We refer to these as the conventional mutations provided by µBERT, denoted by µBERT conv in our evaluation, previously introduced in the preliminary version of the approach [18].…”
Section: Ast Nodes Selectionmentioning
confidence: 99%
“…To answer this question, we generate two sets of mutants using µBERT: 1) the first set using all possible mutations that we denote as µBERT and 2) a second one using only the conventional µBERT' mutations -part of our preliminary implementation [18], excluding the additive ones -that we denote as µBERT conv . Then we evaluate the fault detection ability of test suites selected to kill the mutants from each set.…”
Section: Research Questionsmentioning
confidence: 99%
See 2 more Smart Citations
“…The usage of this model in the fault injection domain shows its ability to seed "natural" faults that, we can say, semantically resemble real faults [34], [35]. Effectively, the faults injected resemble what a real programmer could write (regarding the programmatic rules, convention, etc) [36].…”
Section: Nlp For Fault Injection (Not Vulnerability Injection)mentioning
confidence: 99%