2023
DOI: 10.1109/access.2023.3239668
|View full text |Cite
|
Sign up to set email alerts
|

Membership Inference Attacks With Token-Level Deduplication on Korean Language Models

Abstract: The confidentiality threat against training data has become a significant security problem in neural language models. Recent studies have shown that memorized training data can be extracted by injecting well-chosen prompts into generative language models. While these attacks have achieved remarkable success in the English-based Transformer architecture, it is unclear whether they are still effective in other language domains. This paper studies the effectiveness of attacks against Korean models and the potenti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2025
2025

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(1 citation statement)
references
References 59 publications
0
1
0
Order By: Relevance
“…In an unguided attack, the adversary does not know the sample to be extracted from the model. The adversary simply attempts to extract any training point, contained anywhere in the training corpus [10,12,13,40].…”
Section: Data Extraction Attacksmentioning
confidence: 99%
“…In an unguided attack, the adversary does not know the sample to be extracted from the model. The adversary simply attempts to extract any training point, contained anywhere in the training corpus [10,12,13,40].…”
Section: Data Extraction Attacksmentioning
confidence: 99%