2022
DOI: 10.48550/arxiv.2210.17029
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Poison Attack and Defense on Deep Source Code Processing Models

Abstract: In the software engineering (SE) community, deep learning (DL) has recently been applied to many source code processing tasks, achieving state-of-the-art results. Due to the poor interpretability of DL models, their security vulnerabilities require scrutiny. Recently, researchers have identified an emergent security threat in the DL field, namely poison attack. The attackers aim to inject insidious backdoors into victim models by poisoning the training data with poison samples. Poisoned models work normally wi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
19
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(19 citation statements)
references
References 40 publications
0
19
0
Order By: Relevance
“…Recent work addressed the threat of data poisoning for neural models of source code, i.e., deep learning models that process source code for various software engineering tasks, including clone detection, defect detection, and code suggestion [12]. Wan et al [2] poisoned neural code search systems to manipulate the ranking list of suggested code snippets by injecting backdoors in the training data.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Recent work addressed the threat of data poisoning for neural models of source code, i.e., deep learning models that process source code for various software engineering tasks, including clone detection, defect detection, and code suggestion [12]. Wan et al [2] poisoned neural code search systems to manipulate the ranking list of suggested code snippets by injecting backdoors in the training data.…”
Section: Related Workmentioning
confidence: 99%
“…In backdoor attacks, an attacker's goal is to inject a backdoor into the AI model so that the inputs containing a so-called trigger, i.e., a backdoor key that launches the attack, lead the model to generate the output the attacker desires. Li et al [12] presented both a poison attack framework, named CodePoisoner, and a defense approach, named CodeDetector to deceive deep learning models in defect detection, clone detection and code repair. Ramakrishnan et al [13] made advances in the identification of backdoors, thus enabling the detection of poisoned data.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations