2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00027
|View full text |Cite
|
Sign up to set email alerts
|

Continual Learning with Lifelong Vision Transformer

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
28
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 53 publications
(28 citation statements)
references
References 55 publications
0
28
0
Order By: Relevance
“…Dytox [58] dynamically learns new task tokens, which are then utilized to make the learned embeddings more relevant to the specific task. Lifelong ViT [59] and contrastive ViT [60] introduce crossattention mechanisms between tasks through external key vectors, and they slow down the changes to these keys to mitigate forgetting. Despite the use of complex mechanisms to prevent forgetting, these methods still require fine-tuning of the network for new classes, which can result in interference with previously learned knowledge.…”
Section: Self-supervisedmentioning
confidence: 99%
“…Dytox [58] dynamically learns new task tokens, which are then utilized to make the learned embeddings more relevant to the specific task. Lifelong ViT [59] and contrastive ViT [60] introduce crossattention mechanisms between tasks through external key vectors, and they slow down the changes to these keys to mitigate forgetting. Despite the use of complex mechanisms to prevent forgetting, these methods still require fine-tuning of the network for new classes, which can result in interference with previously learned knowledge.…”
Section: Self-supervisedmentioning
confidence: 99%
“…Continual Learning: CL approaches can be broadly classified into -1) Exemplar-replay methods, 2) regularization methods and 3) dynamic architecture methods. To avoid forgetting when learning a new task, replay approaches repeat past task samples that are kept in raw format [2,4,6,7,9,16,25,40,52] or generated with a generative model [45]. Usually, replay-based approaches have a fixed memory which stores samples.…”
Section: Related Workmentioning
confidence: 99%
“…Recently Dytox [16] proposed to learn new tasks through the expansion of special tokens known as task tokens. Another recent approach, LVT [52], proposed an inter-task attention mechanism that absorbs the previous tasks' information and slows down the drift of information between previous and current tasks. Both Dytox and LVT require extra memory for stor- ing training instances from previous tasks.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations