2022
DOI: 10.1109/tsc.2020.3000900
|View full text |Cite
|
Sign up to set email alerts
|

Backdoor Attacks Against Transfer Learning With Pre-Trained Deep Learning Models

Abstract: Transfer learning provides an effective solution for feasibly and fast customize accurate Student models, by transferring the learned knowledge of pre-trained Teacher models over large datasets via fine-tuning. Many pre-trained Teacher models used in transfer learning are publicly available and maintained by public platforms, increasing their vulnerability to backdoor attacks. In this paper, we demonstrate a backdoor threat to transfer learning tasks on both image and time-series data leveraging the knowledge … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
34
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 69 publications
(34 citation statements)
references
References 37 publications
0
34
0
Order By: Relevance
“…Our code can be accessed at https://github.com/goel96vibhor/AdvWeightPerturbations through data poisoning attacks. Wang et al[18] proposes a backdoor injection scheme to defeating pruning-based, retraining-based and input pre-processing-based defenses. In parallel work, Kurita et al[14] expose the risk of the pre-trained BERT[4] model to backdoor injection attacks mimicking a model capture scenario.…”
mentioning
confidence: 99%
“…Our code can be accessed at https://github.com/goel96vibhor/AdvWeightPerturbations through data poisoning attacks. Wang et al[18] proposes a backdoor injection scheme to defeating pruning-based, retraining-based and input pre-processing-based defenses. In parallel work, Kurita et al[14] expose the risk of the pre-trained BERT[4] model to backdoor injection attacks mimicking a model capture scenario.…”
mentioning
confidence: 99%
“…However, using pre-trained models from foreign sources can pose a risk as the models can be subject to biases and adversarial attacks, as introduced above. For example, pre-trained models may not properly reflect certain environmental constraints or contain backdoors by inserting classification triggers, for example, to misclassify medical images (Wang et al 2020). Governmental interventions to redirect or suppress predictions are conceivable as well.…”
Section: Resource Limitations and Transfer Learningmentioning
confidence: 99%
“…Different from adversarial attacks which usually act during the inference process of a neural model [17,38,49,53,63,63,66,74,84,85], backdoor attacks hack the model during training [10,22,40,51,61,62,75,82]. Defending against such attacks is challenging [8,23,37,41,57,73] because users have no idea of what kinds of poison has been injected into model training.…”
Section: Backdoor Attack and Defensementioning
confidence: 99%