Our new X account is live! Follow @wizwand_team for updates
Home
/
Benchmarks
Downstream Task Evaluation on MNLI, SCIQ, LAMBADA, HellaSwag, ARC, and MMLU
Loading...
0.3852
MNLI Acc
FusedKV
0.3722
0.375575
0.37895
0.382325
Dec 3, 2025
MNLI Acc
SCIQ Acc
LAMBADA Acc
HellaSwag Acc
ARC-E Acc
ARC-C Acc
MMLU Acc
Average Accuracy
Updated 4d ago
Evaluation Results
Method
Method
Links
MNLI Acc
SCIQ Acc
LAMBADA Acc
HellaSwag Acc
ARC-E Acc
ARC-C Acc
MMLU Acc
Average Accuracy
FusedKV
Model Parameters=4B
2025.12
0.3852
0.952
0.5018
0.6994
0.7778
0.4863
0.3983
0.6001
Vanilla
Model Parameters=4B
2025.12
0.3727
0.96
0.496
0.6892
0.7643
0.4659
0.3871
0.5907
Feedback
Search any
task
Search any
task