Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

NIPQ: Noise proxy-based Integrated Pseudo-Quantization

About

Straight-through estimator (STE), which enables the gradient flow over the non-differentiable function via approximation, has been favored in studies related to quantization-aware training (QAT). However, STE incurs unstable convergence during QAT, resulting in notable quality degradation in low precision. Recently, pseudoquantization training has been proposed as an alternative approach to updating the learnable parameters using the pseudo-quantization noise instead of STE. In this study, we propose a novel noise proxy-based integrated pseudoquantization (NIPQ) that enables unified support of pseudoquantization for both activation and weight by integrating the idea of truncation on the pseudo-quantization framework. NIPQ updates all of the quantization parameters (e.g., bit-width and truncation boundary) as well as the network parameters via gradient descent without STE instability. According to our extensive experiments, NIPQ outperforms existing quantization algorithms in various vision and language applications by a large margin.

Juncheol Shin, Junhyuk So, Sein Park, Seungyeop Kang, Sungjoo Yoo, Eunhyeok Park• 2022

Related benchmarks

TaskDatasetResultRank
Natural Language UnderstandingGLUE
SST-291.97
531
Super-ResolutionSet14 (test)
PSNR33.29
254
Super-ResolutionUrban100 (test)
PSNR31.36
220
Super-ResolutionSet5 (test)
PSNR37.74
192
Super-ResolutionBSDS100 (test)
PSNR32.01
97
Image ClassificationImageNet-1k (val)
Top-1 Accuracy69.2
18
Image ClassificationImageNet-1k (val)
Top-1 Accuracy72.3
10
Image ClassificationImageNet (val)
Top-1 Accuracy71.2
9
Showing 8 of 8 rows

Other info

Code

Follow for update