2021
DOI: 10.48550/arxiv.2112.02706
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Achieving Forgetting Prevention and Knowledge Transfer in Continual Learning

Abstract: Continual learning (CL) learns a sequence of tasks incrementally with the goal of achieving two main objectives: overcoming catastrophic forgetting (CF) and encouraging knowledge transfer (KT) across tasks. However, most existing techniques focus only on overcoming CF and have no mechanism to encourage KT, and thus do not do well in KT. Although several papers have tried to deal with both CF and KT, our experiments show that they suffer from serious CF when the tasks do not have much shared knowledge. Another … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
9
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(9 citation statements)
references
References 37 publications
(62 reference statements)
0
9
0
Order By: Relevance
“…Recent years have seen a surge of research efforts focusing on the integration of advanced pre-trained models, such as BERT (Bidirectional Encoder Representations from Transformers) [5], into continual learning approaches. BERT adapters, lightweight task-specific neural network components, have emerged as a promising solution [9], which chooses to freeze the parameters of the pre-trained BERT, only tuning a small number of parameters of injected taskspecific adapters [9][10][11]. A BERT adapter can be a relatively small 2-layer fully connected network or a Capsule Networks (CapsNet) [12,13] that uses vector capsules instead of scalar feature detectors.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Recent years have seen a surge of research efforts focusing on the integration of advanced pre-trained models, such as BERT (Bidirectional Encoder Representations from Transformers) [5], into continual learning approaches. BERT adapters, lightweight task-specific neural network components, have emerged as a promising solution [9], which chooses to freeze the parameters of the pre-trained BERT, only tuning a small number of parameters of injected taskspecific adapters [9][10][11]. A BERT adapter can be a relatively small 2-layer fully connected network or a Capsule Networks (CapsNet) [12,13] that uses vector capsules instead of scalar feature detectors.…”
Section: Related Workmentioning
confidence: 99%
“…Although the CL capability of existing advanced models have been empirically demonstrated they still struggle with CF issues when tasks lack substantial shared knowledge [11,14].…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations