2023
DOI: 10.48550/arxiv.2302.03908
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Syntax and Domain Aware Model for Unsupervised Program Translation

Abstract: There is growing interest in software migration as the development of software and society. Manually migrating projects between languages is error-prone and expensive. In recent years, researchers have begun to explore automatic program translation using supervised deep learning techniques by learning from large-scale parallel code corpus. However, parallel resources are scarce in the programming language domain, and it is costly to collect bilingual data manually. To address this issue, several unsupervised p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 32 publications
(85 reference statements)
0
3
0
Order By: Relevance
“…For the neural code translation task and the neural code repair task, following previous work [32] [6], we use Bilingual Evaluation Understudy (BLEU) score, a common evaluation metric in neural code translation studies [3][3]. BLEU score evaluates the quality of generated functions by measuring the N -gram overlapping between the reference sentence ŷ and the translation y. BLEU-4 is introduced to measure the quality of the translation [2]. The formula is shown in EQ 4.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…For the neural code translation task and the neural code repair task, following previous work [32] [6], we use Bilingual Evaluation Understudy (BLEU) score, a common evaluation metric in neural code translation studies [3][3]. BLEU score evaluates the quality of generated functions by measuring the N -gram overlapping between the reference sentence ŷ and the translation y. BLEU-4 is introduced to measure the quality of the translation [2]. The formula is shown in EQ 4.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…It supports different datasets such as ImageNet, COCO and tasks like image classification and recommendation. At the same time, the software engineering community has established some common downstream tasks (such as program translation [27]) and their corresponding metrics (such as CodeBLEU [45]) for evaluating ML models in this domain. This has facilitated the creation of benchmarks like CodeXGLUE [33] to enable fair and consistent comparison of different ML models.…”
Section: Overview Of Benchmarking ML Toolsmentioning
confidence: 99%
“…Code generation is an essential generation task in the field of natural language processing (NLP) and software engineering [20,24,7,8,18,35,36,16], which deals with automatically generating a piece of executable code from NL utterances. In recent years, a series of Seq2Tree models have made remarkable achievements for code generation [2,38,1,39,27,29,28,33,11,14,43,21]. Specifically, given an NL utterance input, instead of outputting a sequence of code tokens directly, the Seq2Tree model outputs a sequence of AST actions.…”
Section: Introductionmentioning
confidence: 99%