QuEPT: Quantized Elastic Precision Transformers with One-Shot Calibration for Multi-Bit Switching
About
Elastic precision quantization enables multi-bit deployment via a single optimization pass, fitting diverse quantization scenarios.Yet, the high storage and optimization costs associated with the Transformer architecture, research on elastic quantization remains limited, particularly for large language models.This paper proposes QuEPT, an efficient post-training scheme that reconstructs block-wise multi-bit errors with one-shot calibration on a small data slice. It can dynamically adapt to various predefined bit-widths by cascading different low-rank adapters, and supports real-time switching between uniform quantization and mixed precision quantization without repeated optimization. To enhance accuracy and robustness, we introduce Multi-Bit Token Merging (MB-ToMe) to dynamically fuse token features across different bit-widths, improving robustness during bit-width switching. Additionally, we propose Multi-Bit Cascaded Low-Rank adapters (MB-CLoRA) to strengthen correlations between bit-width groups, further improve the overall performance of QuEPT. Extensive experiments demonstrate that QuEPT achieves comparable or better performance to existing state-of-the-art post-training quantization methods.Our code is available at https://github.com/xuke225/QuEPT
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Language Modeling | WikiText2 | Perplexity4.94 | 1875 | |
| Language Modeling | C4 | Perplexity6.53 | 1182 | |
| Visual Question Answering | TextVQA (test) | Accuracy74.1 | 124 | |
| Multimodal Understanding | MMMU (test) | -- | 86 | |
| Visual Question Answering | VizWiz (test) | Accuracy61.3 | 66 | |
| Optical Character Recognition | OCRBench (test) | Score61.2 | 34 | |
| General Language Evaluation | 5 Datasets Zero-shot | Average Accuracy72.87 | 33 | |
| Image Classification | ImageNet-1k (val) | Top-1 Acc (W6A6)83.8 | 23 |