2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00907
|View full text |Cite
|
Sign up to set email alerts
|

DyTox: Transformers for Continual Learning with DYnamic TOken eXpansion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
77
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 172 publications
(79 citation statements)
references
References 18 publications
1
77
0
1
Order By: Relevance
“…The results in Table 1 compare Continual-with other baselines on CIFAR-100 dataset in three different settings (10, 20 and 50 steps). In 10 steps setting, even without any training or fine-tuning, Continual-CLIP achieves competitive results in terms average and last accuracy, compared with the recent state-of-the-art methods such as DyTox (Douillard et al, 2022) and DER (Yan et al, 2021). Specifically, In 20 steps setting, Continual-CLIP reaches 75.95% in "Avg" accuracy, and for the 50 steps setting, it reaches 76.49% in "Avg" accuracy.…”
Section: Resultsmentioning
confidence: 91%
See 3 more Smart Citations
“…The results in Table 1 compare Continual-with other baselines on CIFAR-100 dataset in three different settings (10, 20 and 50 steps). In 10 steps setting, even without any training or fine-tuning, Continual-CLIP achieves competitive results in terms average and last accuracy, compared with the recent state-of-the-art methods such as DyTox (Douillard et al, 2022) and DER (Yan et al, 2021). Specifically, In 20 steps setting, Continual-CLIP reaches 75.95% in "Avg" accuracy, and for the 50 steps setting, it reaches 76.49% in "Avg" accuracy.…”
Section: Resultsmentioning
confidence: 91%
“…For class-incremental settings, we evaluate Continual-CLIP on CIFAR-100, ImageNet-100 & 1K, TinyImageNet under different class splits. (a) In CIFAR-100, we compare the performance on 10 steps (10 new classes per step), 20 steps (5 new classes per step), and 50 steps (2 new classes per step) (Douillard et al, 2022;Yan et al, 2021). (b) In ImageNet-100, we consider two evaluation settings; ImageNet-100-B0 which has the same number of classes for all the steps (i.e., 10 classes per step) and ImageNet-100-B50 that contains 50 classes for the first step and the rest of the 50 classes are observed incrementally in the next 10 steps (5 classes per steps) (Yan et al, 2021).…”
Section: Experimental Protocolsmentioning
confidence: 99%
See 2 more Smart Citations
“…However, including extra data into the current task introduces excessive training time (De Lange et al, 2021). Expansion-based methods (Yoon et al, 2017;2019;Douillard et al, 2022) dynamically allocate new parameters or modules to learn new tasks. While these methods face capacity explosion inevitably after learning a long sequence of tasks.…”
Section: Other Methodsmentioning
confidence: 99%