2021
DOI: 10.48550/arxiv.2102.08355
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Adversarial Targeted Forgetting in Regularization and Generative Based Continual Learning Models

Abstract: Continual (or "incremental") learning approaches are employed when additional knowledge or tasks need to be learned from subsequent batches or from streaming data. However these approaches are typically adversary agnostic, i.e., they do not consider the possibility of a malicious attack. In our prior work, we explored the vulnerabilities of Elastic Weight Consolidation (EWC) to the perceptible misinformation. We now explore the vulnerabilities of other regularization-based as well as generative replay-based co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 20 publications
0
4
0
Order By: Relevance
“…Without any loss of generality, we assume that the target task is Task 1. To degrade the test time performance of the target task, the attacker can insert malicious samples into the training data of all non-target tasks (i.e., all tasks except Task 1 as done in our prior work [18]. However, in order to make the attack even more difficult to detect, we now restrict the attacker to add small amount of imperceptible misinformation only at the last task to degrade the test time performance of the first task.…”
Section: Experiments and Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…Without any loss of generality, we assume that the target task is Task 1. To degrade the test time performance of the target task, the attacker can insert malicious samples into the training data of all non-target tasks (i.e., all tasks except Task 1 as done in our prior work [18]. However, in order to make the attack even more difficult to detect, we now restrict the attacker to add small amount of imperceptible misinformation only at the last task to degrade the test time performance of the first task.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…In preliminary stages of this work, we discussed vulnerability of importance based domain adaptation to poisoning attacks in [16], [17], the vulnerability of EWC under task incremental setting to perceptible backdoor attacks in [7], and the vulnerabilities of regularization and generative replay approaches considering only the simple MNIST dataset and where the attack was launched on all tasks in [18]. This effort significantly expands the prior work to not only the more challenging SVHN and CIFAR10 datasets, but also to a more realistic scenario where the attacker inserts imperceptible misinformation to only a single task (as opposed to all tasks in prior work) of its own choosing.…”
Section: A Regularization and Replay-based Continual Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…While several CL approaches have been proposed to avoid the problem of catastrophic forgetting, but recently it has been found that these approaches are extremely vulnerable to adversarial backdoor attacks [3]- [5], where an intelligent adversary can easily insert miniature amount of misinformation in the training data to deliberately or intentionally disturb the balance between stability and plasticity acquired by the CL model. More specifically, the goal of such an attack is to artificially increase the forgetting of the CL model on a explicitly targeted previously learned task.…”
Section: Introductionmentioning
confidence: 99%