Our new X account is live! Follow @wizwand_team for updates
Home
/
Benchmarks
Language Understanding on HellaSwag, PIQA, MMLU Suite
Loading...
49
HellaSwag Accuracy
OpenLLaMA (V1)
33.4
37.45
41.5
45.55
Dec 28, 2023
HellaSwag Accuracy
PIQA Accuracy
MMLU Accuracy
Average Score
Updated 3d ago
Evaluation Results
Method
Method
Links
HellaSwag Accuracy
PIQA Accuracy
MMLU Accuracy
Average Score
OpenLLaMA (V1)
Parameters=3B
2023.12
49
75
26.95
49.9
INCITE (V1)
Parameters=3B
2023.12
48
74
26.75
48.48
MobileLLaMA
Parameters=2.7B
2023.12
48
75
27.3
49.23
MobileLLaMA
Parameters=1.4B
2023.12
43
73
24.97
44.9
OPT
Parameters=1.3B
2023.12
41
71
24.61
44.66
TinyLLaMA (2T)
Parameters=1.1B
2023.12
40
70
25.41
44.84
Pythia
Parameters=1.4B
2023.12
40
71
25.68
45.77
Galactica
Parameters=1.3B
2023.12
34
63
26.75
44.88
Feedback
Search any
task
Search any
task