2022
DOI: 10.1007/978-3-031-20050-2_40
|View full text |Cite
|
Sign up to set email alerts
|

Transfer Without Forgetting

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(10 citation statements)
references
References 33 publications
0
10
0
Order By: Relevance
“…To assess our proposal, we perform a suite of experiments through the Mammoth framework [12,17,9,6,27,13,14,50], an open-source codebase introduced in [15] for testing CL algorithms. In particular, we show that our method can be easily applied to state-of-the-art replay methods and enhance their performance in a wide variety of challenging settings and backbone architectures.…”
Section: Methodsmentioning
confidence: 99%
“…To assess our proposal, we perform a suite of experiments through the Mammoth framework [12,17,9,6,27,13,14,50], an open-source codebase introduced in [15] for testing CL algorithms. In particular, we show that our method can be easily applied to state-of-the-art replay methods and enhance their performance in a wide variety of challenging settings and backbone architectures.…”
Section: Methodsmentioning
confidence: 99%
“…In this section, we provide comprehensive per-step comparative results of our Baseline and Baseline++, juxtaposed with the adapted state-of-the-art methods on two task splits (two-step and five-step) of CIFAR-10 (C10) [28], CIFAR-100 (C100) [28], TinyImageNet-200 (T200) [29], 3 https://github.com/aimagelab/mammoth CUB-200 (B200) [42] and Herbarium-683 (H683) [38] As depicted in the reported figures, the overall accuracy exhibits a decline as the task sequence progresses, whereas the maximum forgetting for the novel classes discovered during the first step experiences an increase, attributable to the catastrophic forgetting issue [41]. In the context of longer task sequences (five-step split, as observed in the top half of the figures), the forgetting issue is exacerbated due to more frequent model updates.…”
Section: Detailed Experimental Resultsmentioning
confidence: 99%
“…All adapted methods unfreeze only the last transformer block of the feature extractor, except ResTune that unfreezes the last two blocks. This is because it is unnecessary to unlock all blocks of the largescale pre-trained model, as previously observed in [45] and [3]. Additional implementation details of the adapted methods are provided in the supplementary material.…”
Section: Comparison With the State-of-the-art Methodsmentioning
confidence: 99%
See 2 more Smart Citations