Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learned Step Size Quantization

About

Deep networks run with low precision operations at inference time offer power and space advantages over high precision alternatives, but need to overcome the challenge of maintaining high accuracy as precision decreases. Here, we present a method for training such networks, Learned Step Size Quantization, that achieves the highest accuracy to date on the ImageNet dataset when using models, from a variety of architectures, with weights and activations quantized to 2-, 3- or 4-bits of precision, and that can train 3-bit models that reach full precision baseline accuracy. Our approach builds upon existing methods for learning weights in quantized networks by improving how the quantizer itself is configured. Specifically, we introduce a novel means to estimate and scale the task loss gradient at each weight and activation layer's quantizer step size, such that it can be learned in conjunction with other network parameters. This approach works using different levels of precision as needed for a given system and requires only a simple modification of existing training code.

Steven K. Esser, Jeffrey L. McKinstry, Deepika Bablani, Rathinakumar Appuswamy, Dharmendra S. Modha• 2019

Related benchmarks

TaskDatasetResultRank
Object DetectionCOCO 2017 (val)
AP33.9
2454
Language ModelingWikiText-2 (test)
PPL16.75
1541
Image ClassificationImageNet-1k (val)
Top-1 Accuracy76.7
1453
Image ClassificationImageNet (val)
Top-1 Acc78.5
1206
Visual Question AnsweringVQA v2
Accuracy77.3
1165
Visual Question AnsweringVizWiz
Accuracy49.8
1043
Object Hallucination EvaluationPOPE
Accuracy85
935
Object DetectionPASCAL VOC 2007 (test)
mAP49.7
821
Language ModelingWikiText-103 (test)
Perplexity15.24
524
Text-based Visual Question AnsweringTextVQA
Accuracy57.3
496
Showing 10 of 32 rows

Other info

Follow for update