2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00329
|View full text |Cite
|
Sign up to set email alerts
|

Unifying Heterogeneous Classifiers With Distillation

Abstract: In this paper, we study the problem of unifying knowledge from a set of classifiers with different architectures and target classes into a single classifier, given only a generic set of unlabelled data. We call this problem Unifying Heterogeneous Classifiers (UHC). This problem is motivated by scenarios where data is collected from multiple sources, but the sources cannot share their data, e.g., due to privacy concerns, and only privately trained models can be shared. In addition, each source may not be able t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
32
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 43 publications
(32 citation statements)
references
References 32 publications
0
32
0
Order By: Relevance
“…Similar limits can be found in recent classifier amalgamation works 1 . A few recent works [21] [22] [23] has been proposed to unifying heterogeneous teacher classifiers. Without a predefined dustbin class, [23] requires overlapped classes of objects recognized by teacher models, otherwise the model failed to find an optimal feature 1 A detailed comparison at https://github.com/zju-vipa/KamalEngine alignments.…”
Section: B Multi-teacher Knowledge Distillationmentioning
confidence: 99%
See 2 more Smart Citations
“…Similar limits can be found in recent classifier amalgamation works 1 . A few recent works [21] [22] [23] has been proposed to unifying heterogeneous teacher classifiers. Without a predefined dustbin class, [23] requires overlapped classes of objects recognized by teacher models, otherwise the model failed to find an optimal feature 1 A detailed comparison at https://github.com/zju-vipa/KamalEngine alignments.…”
Section: B Multi-teacher Knowledge Distillationmentioning
confidence: 99%
“…where q clo pc k P C´iq is estimated by the sum of the dustbin probabilities. As discussed in [21], grouping classes in C´i into a single dustbin classes imposes design constraint on f i . We argue that the presence of dustbin facilitates efficient implementation of the distillation loss function.…”
Section: A Classification and Conventional Distillation Lossesmentioning
confidence: 99%
See 1 more Smart Citation
“…[24] applies multi-teacher learning into multi-task learning where each teacher corresponds to a task. Similarly, [26] trains a classifier in each source and unifying their classifications on an integrated label space. Nevertheless, the multi-domain or multi-task setting limits the application of the method to more general scenarios for a single task in a single domain.…”
Section: Related Workmentioning
confidence: 99%
“…Some recent studies [28,6] have taken a step along this direction by either simply treating different teachers equally or manually tuning the teacher importance weights, lacking the ability of automatically discriminating different teachers. Some other studies [19,24,26] have applied multiple teacher networks to multi-task or multidomain settings. They assume each task or domain corresponds to a teacher network.…”
Section: Introductionmentioning
confidence: 99%