Proceedings of the 39th IEEE/ACM International Conference on Automated Software Engineering 2024
DOI: 10.1145/3691620.3695480
|View full text |Cite
|
Sign up to set email alerts
|

RMCBench: Benchmarking Large Language Models' Resistance to Malicious Code

Jiachi Chen,
Qingyuan Zhong,
Yanlin Wang
et al.

Abstract: that current LLMs have a limited ability to resist malicious code generation with an average refusal rate of 40.36% in text-to-code scenario and 11.52% in code-to-code scenario. The average refusal rate of all LLMs in RMCBench is only 28.71%; ChatGPT-4 has a refusal rate of only 35.73%. We also analyze the factors that affect LLM's ability to resist malicious code generation and provide implications for developers to enhance model robustness. CCS CONCEPTS• Security and privacy → Software security engineering.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 11 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?