Share your thoughts, 1 month free Claude Pro on us
See more
Home
/
Benchmarks
Large Language Model Evaluation on LLaMA-2 13B
Loading...
4.88
Perplexity
baseline
-522.9248
3,039.7576
6,602.44
10,165.1224
Mar 9, 2026
Perplexity
Zero-Shot Performance
MMLU Score
Updated 1mo ago
Evaluation Results
Method
Method
Links
Perplexity
Zero-Shot Performance
MMLU Score
baseline
#Bits=FP16, #Eff. (w-b...
2026.03
4.88
66.53
52.04
SpinQuant
#Bits=W4A4, #Eff. (w-b...
2026.03
5.2
64.8
47.8
SERQ
#Bits=W4A4, #Eff. (w-b...
2026.03
5.2
64.82
47.17
QuaRot
#Bits=W4A4, #Eff. (w-b...
2026.03
5.41
62.55
47.25
SmoothQ(g128)
#Bits=W4A4, #Eff. (w-b...
2026.03
6.31
61.28
39.83
SmoothQ
#Bits=W4A4, #Eff. (w-b...
2026.03
13,200
34.53
23.85
Feedback
Search any
task
Search any
task