Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

PTQ4ViT: Post-training quantization for vision transformers with twin uniform quantization

About

Quantization is one of the most effective methods to compress neural networks, which has achieved great success on convolutional neural networks (CNNs). Recently, vision transformers have demonstrated great potential in computer vision. However, previous post-training quantization methods performed not well on vision transformer, resulting in more than 1% accuracy drop even in 8-bit quantization. Therefore, we analyze the problems of quantization on vision transformers. We observe the distributions of activation values after softmax and GELU functions are quite different from the Gaussian distribution. We also observe that common quantization metrics, such as MSE and cosine distance, are inaccurate to determine the optimal scaling factor. In this paper, we propose the twin uniform quantization method to reduce the quantization error on these activation values. And we propose to use a Hessian guided metric to evaluate different scaling factors, which improves the accuracy of calibration at a small cost. To enable the fast quantization of vision transformers, we develop an efficient framework, PTQ4ViT. Experiments show the quantized vision transformers achieve near-lossless prediction accuracy (less than 0.5% drop at 8-bit quantization) on the ImageNet classification task.

Zhihang Yuan, Chenhao Xue, Yiqi Chen, Qiang Wu, Guangyu Sun• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet-1K
Top-1 Acc83
1239
Image Super-resolutionManga109
PSNR37.59
821
Image Super-resolutionSet5
PSNR37.43
692
Image Super-resolutionSet14
PSNR33.19
506
Single Image Super-ResolutionUrban100
PSNR31.54
500
Image Super-resolutionUrban100
PSNR27.43
406
Image ClassificationImageNet (val)--
300
ClassificationImageNet1K
Accuracy82.97
202
Text-to-Image RetrievalMS-COCO
R@151.62
151
Image-to-Text RetrievalMS-COCO
R@169.74
132
Showing 10 of 18 rows

Other info

Follow for update