Proceedings of the 1st Workshop on Natural Language Processing for Programming (NLP4Prog 2021) 2021
DOI: 10.18653/v1/2021.nlp4prog-1.6
|View full text |Cite
|
Sign up to set email alerts
|

DIRECT : A Transformer-based Model for Decompiled Identifier Renaming

Abstract: Decompiling binary executables to high-level code is an important step in reverse engineering scenarios, such as malware analysis and legacy code maintenance. However, the generated high-level code is difficult to understand since the original variable names are lost. In this paper, we leverage transformer models to reconstruct the original variable names from decompiled code. Inherent differences between code and natural language present certain challenges in applying conventional transformer-based architectu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(1 citation statement)
references
References 17 publications
0
1
0
Order By: Relevance
“…Code-to-Code Deobfuscation DeGuard (Bichsel et al, 2016), Autonym (Vasilescu et al, 2017), Debin , JS-Neat DIRE (Lacomis et al, 2019), Artuso et al (2021) VarBERT (Banerjee et al, 2021), DIRECT (Nitin et al, 2021), SLaDe (Armengol-Estapé et al, 2023), LmPa (Xu et al, 2023b) Figure 5: Evaluation tasks for code processing: models and methods for code-to-code tasks, part 2. For each task the first column is non-neural methods (mostly n-gram or TF-IDF based methods, but also including other heuristics based methods); the second column is non-Transformer neural methods; the third column is Transformer based methods.…”
Section: Code Evaluationmentioning
confidence: 99%
“…Code-to-Code Deobfuscation DeGuard (Bichsel et al, 2016), Autonym (Vasilescu et al, 2017), Debin , JS-Neat DIRE (Lacomis et al, 2019), Artuso et al (2021) VarBERT (Banerjee et al, 2021), DIRECT (Nitin et al, 2021), SLaDe (Armengol-Estapé et al, 2023), LmPa (Xu et al, 2023b) Figure 5: Evaluation tasks for code processing: models and methods for code-to-code tasks, part 2. For each task the first column is non-neural methods (mostly n-gram or TF-IDF based methods, but also including other heuristics based methods); the second column is non-Transformer neural methods; the third column is Transformer based methods.…”
Section: Code Evaluationmentioning
confidence: 99%