2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.00164
|View full text |Cite
|
Sign up to set email alerts
|

On Learning the Geodesic Path for Incremental Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
50
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 88 publications
(50 citation statements)
references
References 25 publications
0
50
0
Order By: Relevance
“…Aleatoric uncertainty and self-attention are involved in the proposed distillation loss. Simon et al [55] proposed to conduct knowledge distillation on the low-dimensional manifolds between the model outputs of previous and new tasks, which was shown to better mitigate catastrophic forgetting. Hu et al [56] found that the knowledge distillation based approaches do not have a consistent causal effect compared to end-to-end feature learning.…”
Section: Distillation In the Presence Of Sufficient Data Instancesmentioning
confidence: 99%
“…Aleatoric uncertainty and self-attention are involved in the proposed distillation loss. Simon et al [55] proposed to conduct knowledge distillation on the low-dimensional manifolds between the model outputs of previous and new tasks, which was shown to better mitigate catastrophic forgetting. Hu et al [56] found that the knowledge distillation based approaches do not have a consistent causal effect compared to end-to-end feature learning.…”
Section: Distillation In the Presence Of Sufficient Data Instancesmentioning
confidence: 99%
“…[17] uses the causal effect on knowledge distillation to rectify class imbalance. [43] introduces geodesic path to traditional knowledge distillation. [1] combines task-wise knowledge distillation and separated softmax for bias compensation.…”
Section: Related Workmentioning
confidence: 99%
“…In the standard class-incremental learning [37,41,43], there are a sequence of streaming tasks T = {T t } T t=1 , where T denotes the task number, and the t-th task T t = {x t i , y t i } N t i=1 consists of N t pairs of samples x t i and their one-hot encoding labels y t i ∈ Y t . Y t represents the label space of the t-th task including C t new classes that are different from…”
Section: Problem Definitionmentioning
confidence: 99%
“…Incremental learning: Incremental learning means learning from a sequence of data which appear over time. In the literature [33], incremental learning techniques are categorized into three groups, task-incremental learning [4,28,21], domain-incremental learning [39,29], and classincremental learning [26,1,13,36,31]. In this paper, we are only concerned with the third category, class-incremental learning, as we consider a unified output where the task label is not available during test time.…”
Section: Related Workmentioning
confidence: 99%
“…Wu et al [36] proposed a method for large scale incremental learning, where they correct the bias in the output of the model with the help of a linear model. Simon et al [31] propose a novel approach to the arsenal of distillation techniques. They construct lowdimensional manifolds for previous and current responses and minimize the dissimilarity between the geodesic responses connecting the manifolds.…”
Section: Related Workmentioning
confidence: 99%