Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: Industry Track 2022
DOI: 10.18653/v1/2022.emnlp-industry.41
|View full text |Cite
|
Sign up to set email alerts
|

Fast Vocabulary Transfer for Language Model Compression

Abstract: Real-world business applications require a trade-off between language model performance and size. We propose a new method for model compression that relies on vocabulary transfer. We evaluate the method on various vertical domains and downstream tasks. Our results indicate that vocabulary transfer can be effectively used in combination with other compression techniques, yielding a significant reduction in model size and inference time while marginally compromising on performance.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 19 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?