2021
DOI: 10.48550/arxiv.2104.14696
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer

Abstract: Recent applications pose requirements of both cross-domain knowledge transfer and model compression to machine learning models due to insufficient training data and limited computational resources. In this paper, we propose a new knowledge distillation model, named Spirit Distillation (SD), which is a model compression method with multidomain knowledge transfer. The compact student network mimics out a representation equivalent to the front part of the teacher network, through which the general knowledge can b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 30 publications
0
1
0
Order By: Relevance
“…Our framework also intends to optimize the memory storage of each new domain, where we maximally reuse the parameters of the base generative model and only save the specific style modulation parameters (≈ 10MB) instead of the whole generative model (≈ 156MB). With recent advances in model compression and pruning [42], we believe the memory consumption could be further reduced in future.…”
Section: Discussionmentioning
confidence: 99%
“…Our framework also intends to optimize the memory storage of each new domain, where we maximally reuse the parameters of the base generative model and only save the specific style modulation parameters (≈ 10MB) instead of the whole generative model (≈ 156MB). With recent advances in model compression and pruning [42], we believe the memory consumption could be further reduced in future.…”
Section: Discussionmentioning
confidence: 99%