2022
DOI: 10.48550/arxiv.2208.00777
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

$\textrm{D}^3\textrm{Former}$: Debiased Dual Distilled Transformer for Incremental Learning

Abstract: In class incremental learning (CIL) setting, groups of classes are introduced to a model in each learning phase. The goal is to learn a unified model performant on all the classes observed so far. Given the recent popularity of Vision Transformers (ViTs) in conventional classification settings, an interesting question is to study their continual learning behaviour. In this work, we develop a Debiased Dual Distilled Transformer for CIL dubbed D 3 Former. The proposed model leverages a hybrid nested ViT design t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 33 publications
(72 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?