2020
DOI: 10.48550/arxiv.2003.03603
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generative Low-bitwidth Data Free Quantization

Abstract: Neural network quantization is an effective way to compress deep models and improve the execution latency and energy efficiency, so that they can be deployed on mobile or embedded devices. Existing quantization methods require original data for calibration or fine-tuning to get better performance. However, in many real-world scenarios, the data may not be available due to confidential or private issues, making existing quantization methods not applicable. Moreover, due to the absence of original data, the rece… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 46 publications
(46 reference statements)
0
3
0
Order By: Relevance
“…Generative adversarial networks (GAN) [2,33] can generate images with high fidelity, but still need real images as a reference while training the generator, which is not fully data-free. Recently, Chen et al [5] and Xu et al [28] proposed to use a generator to synthesize images from a pre-trained model and simultaneously train the student network. Further, Yin et al [29] proposed to synthesize images from the pre-trained teacher network using regularization terms and Jensen-Shannon divergence loss.…”
Section: Related Workmentioning
confidence: 99%
“…Generative adversarial networks (GAN) [2,33] can generate images with high fidelity, but still need real images as a reference while training the generator, which is not fully data-free. Recently, Chen et al [5] and Xu et al [28] proposed to use a generator to synthesize images from a pre-trained model and simultaneously train the student network. Further, Yin et al [29] proposed to synthesize images from the pre-trained teacher network using regularization terms and Jensen-Shannon divergence loss.…”
Section: Related Workmentioning
confidence: 99%
“…DFC and ZeroQ proposed to synthesize calibration data by gradient descent under the supervision of BatchNorm statistics [10,3]. Generative Adversarial Networks (GAN) is also utilized to generate synthetic calibration data [37], in which the generator generates calibration data and the discriminator is the quantized model. One problem with these optimization-based methods is the high time complexity.…”
Section: Post-training Quantizationmentioning
confidence: 99%
“…ZeroQ showed its arXiv:2105.07331v1 [cs.LG] 16 May 2021 effectiveness on various datasets including ImageNet [7] and MSCOCO [23]. Another method is to use Generative Adversarial Networks (GAN) to create fake calibration data [37]. The proposed framework named GDFQ is trained with BatchNorm statistics loss, Cross Entropy classification loss and knowledge distillation loss.…”
Section: Introductionmentioning
confidence: 99%