Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence 2022
DOI: 10.24963/ijcai.2022/164
|View full text |Cite
|
Sign up to set email alerts
|

FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer

Abstract: Generative Adversarial Networks(GANs) are powerful generative models on numerous tasks and datasets but are also known for their training instability and mode collapse. The latter is because the optimal transportation map is discontinuous, but DNNs can only approximate continuous ones. One way to solve the problem is to introduce multiple discriminators or generators. However, their impacts are limited because the cost function of each component is the same. That is, they are homogeneous. In contrast, multiple… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
54
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 67 publications
(54 citation statements)
references
References 4 publications
0
54
0
Order By: Relevance
“…The transformed formulas () and () still include hardware‐unfriendly exponents and divisions that would consume many hardware resources if using 32‐bit single‐precision floating‐point (FP32) directly. Since transformer is less affected by low‐precision calculations [11, 12], we use 16‐bit fixed‐point (INT16) to perform all softmax and GELU operations. Next, we optimize the hardware designs for softmax and GELU.…”
Section: Transformations Of Softmax and Gelumentioning
confidence: 99%
“…The transformed formulas () and () still include hardware‐unfriendly exponents and divisions that would consume many hardware resources if using 32‐bit single‐precision floating‐point (FP32) directly. Since transformer is less affected by low‐precision calculations [11, 12], we use 16‐bit fixed‐point (INT16) to perform all softmax and GELU operations. Next, we optimize the hardware designs for softmax and GELU.…”
Section: Transformations Of Softmax and Gelumentioning
confidence: 99%
“…Zhu et al (2021b) identifies the importance of different dimensions in each layer of ViTs and then executes model pruning. Liu et al (2021b); Lin et al (2022); Li et al (2022d) quantize weights and inputs to compress the learning model. Li et al (2022a) studies automated progressive learning that automatically increases the model capacity onthe-fly.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Unfortunately, to the best of our knowledge, we are not aware of any open-resource PTQ method specifically for LIC models. For fair comparison, we implement the Range-Adaptive Quantization (RAQ) (Hong et al, 3615G% /X>)3@ /X>,17@2XUV /X>,17@5$4 /X>,17@)49L7 &KHQJ>)3@ &KHQJ>,17@2XUV 0LQQHQ>)3@ 0LQQHQ>,17@2XUV 970 %3* (a) %LWUDWHESS 3615G% /X>)3@ /X>,17@2XUV /X>,17@5$4 /X>,17@)49L7 &KHQJ>)3@ &KHQJ>,17@2XUV 0LQQHQ>)3@ 0LQQHQ>,17@2XUV 970 %3* (b) 2020) originally requiring model retraining as a PTQ approach; On the other hand, we also include the FQ-ViT (Lin et al, 2022) for comparative study. It is a PTQ method originally designed for image classification and objective detection using Transformer backbone.…”
Section: Comparison Setupmentioning
confidence: 99%