Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

BitNet: Scaling 1-bit Transformers for Large Language Models

About

The increasing size of large language models has posed challenges for deployment and raised concerns about environmental impact due to high energy consumption. In this work, we introduce BitNet, a scalable and stable 1-bit Transformer architecture designed for large language models. Specifically, we introduce BitLinear as a drop-in replacement of the nn.Linear layer in order to train 1-bit weights from scratch. Experimental results on language modeling show that BitNet achieves competitive performance while substantially reducing memory footprint and energy consumption, compared to state-of-the-art 8-bit quantization methods and FP16 Transformer baselines. Furthermore, BitNet exhibits a scaling law akin to full-precision Transformers, suggesting its potential for effective scaling to even larger language models while maintaining efficiency and performance benefits.

Hongyu Wang, Shuming Ma, Li Dong, Shaohan Huang, Huaijie Wang, Lingxiao Ma, Fan Yang, Ruiping Wang, Yi Wu, Furu Wei• 2023

Related benchmarks

TaskDatasetResultRank
Language ModelingWikiText2
Perplexity11.2
2839
Language ModelingC4
Perplexity11.17
1422
Commonsense ReasoningCommonsenseQA
Accuracy (pass@1)47.2
45
Zero-shot ReasoningReasoning Suite (ARC-e, ARC-c, HellaSwag, PIQA, Winogrande) zero-shot
ARC-e Accuracy0.5332
41
Large Language Model InferenceDecode Phase BS=1
Latency (s)0.158
18
Commonsense Question AnsweringCommonsense QA
BoolQ Accuracy62
17
Large Language Model InferencePrefill Phase SeqLen=2k
Prefill Time (s)0.026
15
Showing 7 of 7 rows

Other info

Follow for update