Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Achieving binary weight and activation for LLMs using Post-Training Quantization

About

Quantizing large language models (LLMs) to 1-bit precision significantly reduces computational costs, but existing quantization techniques suffer from noticeable performance degradation when using weight and activation precisions below 4 bits (W4A4). In this paper, we propose a post-training quantization framework with W(1+1)A(1*4) configuration, where weights are quantized to 1 bit with an additional 1 bit for fine-grain grouping and activations are quantized to 1 bit with a 4-fold increase in the number of channels. For weight quantization, we propose utilizing Hessian-aware fine-grained grouping along with an EM-based quantization scheme. For activation quantization, we decompose INT4-quantized activations into a 4 * INT1 format equivalently and simultaneously smooth the scaling factors based on quantization errors, which further reduces the quantization errors in activations. Our method surpasses state-of-the-art (SOTA) LLM quantization baselines on W2A4 across multiple tasks, pushing the boundaries of existing LLM quantization methods toward fully binarized models. Code is available at https://github.com/JimmyCrave/LLM-PTQ-binarization.

Siqing Song, Chuang Wang, Ruiqi Wang, Yi Yang, Xu-Yao Zhang• 2025

Related benchmarks

TaskDatasetResultRank
Language ModelingWikiText2
Perplexity7.17
1875
Language ModelingC4
Perplexity10.18
1182
Language ModelingPTB
Perplexity37.2
650
Language UnderstandingMMLU (test)
MMLU Average Accuracy28
136
Question AnsweringQA Suite Zero-shot (PIQA, ARC-E, ARC-C, BoolQ, HellaSwag, WinoGrande)
PIQA Accuracy72.09
47
Showing 5 of 5 rows

Other info

Follow for update