2020 International Joint Conference on Neural Networks (IJCNN) 2020
DOI: 10.1109/ijcnn48605.2020.9206809
|View full text |Cite
|
Sign up to set email alerts
|

Targeted Forgetting and False Memory Formation in Continual Learners through Adversarial Backdoor Attacks

Abstract: In this brief, we show that sequentially learning new information presented to a continual (incremental) learning model introduces new security risks: an intelligent adversary can introduce small amount of misinformation to the model during training to cause deliberate forgetting of a specific task or class at test time, thus creating "false memory" about that task. We demonstrate such an adversary's ability to assume control of the model by injecting "backdoor" attack samples to commonly used generative repla… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1
1

Relationship

2
6

Authors

Journals

citations
Cited by 11 publications
(6 citation statements)
references
References 29 publications
0
5
0
Order By: Relevance
“…The Fisher information on the pre-trained model is considered as the cause of this problem; it contains information to be forgotten. We believe that the reason for the insufficient accuracy is that the effects of the information to be forgotten contained in Fisher information hardly disappear in EWC (Umer, Dawson, and Polikar 2020).…”
Section: Discussionmentioning
confidence: 95%
“…The Fisher information on the pre-trained model is considered as the cause of this problem; it contains information to be forgotten. We believe that the reason for the insufficient accuracy is that the effects of the information to be forgotten contained in Fisher information hardly disappear in EWC (Umer, Dawson, and Polikar 2020).…”
Section: Discussionmentioning
confidence: 95%
“…In the literature, researchers have used incremental learning techniques to detect fake news using Artificial Neural Networks (ANN). However, ANNs suffer from catastrophic forgetting which lowers the performance of the model as data streams arrive [20]. The deep learning and neural network-based techniques can classify short text appearing sequentially but require large memory space and training time, thus reducing the performance of the model [19].…”
Section: A Incremental Learning Approachmentioning
confidence: 99%
“…Both regularization-and generative replay-based CL approaches work reasonably well in retaining prior knowledge, but their vulnerability to adversarial attacks has only recently started to be explored. For example, the vulnerability of EWC to perceptible backdoor attacks was investigated in [13], and the vulnerability of related importance based domain adaptation approaches to optimization based poisoning attacks has been discussed in our prior work [14], [15]. However, to the best of our knowledge, the vulnerabilities and robustness of other regularization-based approaches, as well as those of the more successful and robust generative replay-based approaches [5] have not yet been determined.…”
Section: A Continual Learningmentioning
confidence: 99%
“…In our prior work, we showed the vulnerability of a specific regularization-based continual learning algorithm, elastic weight consolidation, to adversarial backdoor poisoning attacks with visibly obvious backdoor patterns (such as small subimages) that are embedded into the training dataset with a false label of the attacker's choosing [6]. In this effort, we extend that work to demonstrate the same vulnerability in other state-of-the-art regularization-based and generative replay-based CL algorithms.…”
Section: Introductionmentioning
confidence: 99%