2022
DOI: 10.1007/s10664-022-10122-9
|View full text |Cite
|
Sign up to set email alerts
|

Software system comparison with semantic source code embeddings

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 59 publications
0
2
0
Order By: Relevance
“…Kovačević et al [46] conducted experiments with the Code2Vec, Code2Seq, and CuBERT models to represent Java methods or classes as code embeddings, facilitating machine-learning-based detection of two code smells, i.e., long method and god class, while Ma et al [26] leveraged the CodeT5, CodeGPT, and CodeBERT models to detect the feature envy code smell. To compare software systems, Karakatič et al [47] utilized a pre-trained Code2Vec model to embed Java methods. The work of Fatima et al [27] employed the CodeBERT model to represent Java test cases, assisting in the prediction of flaky (i.e., non-deterministic) test cases.…”
Section: Pre-trained Models In Code-related Tasksmentioning
confidence: 99%
“…Kovačević et al [46] conducted experiments with the Code2Vec, Code2Seq, and CuBERT models to represent Java methods or classes as code embeddings, facilitating machine-learning-based detection of two code smells, i.e., long method and god class, while Ma et al [26] leveraged the CodeT5, CodeGPT, and CodeBERT models to detect the feature envy code smell. To compare software systems, Karakatič et al [47] utilized a pre-trained Code2Vec model to embed Java methods. The work of Fatima et al [27] employed the CodeBERT model to represent Java test cases, assisting in the prediction of flaky (i.e., non-deterministic) test cases.…”
Section: Pre-trained Models In Code-related Tasksmentioning
confidence: 99%
“…Karakatic et al [92] introduced a novel method for comparing software systems by computing the robust Hausdorff distance between semantic source code embeddings of each program component. The authors utilized a pre-trained neural network model, code2vec, to generate source code vector representations from various open-source libraries.…”
Section: Duplicate Code Detection and Similaritymentioning
confidence: 99%