Our new X account is live! Follow @wizwand_team for updates
Home
/
Benchmarks
Summarization Evaluation on SummEval (Error Metrics)
Loading...
0.495
MSE
MILE-RefHumEval
0.33588
1.40994
2.484
3.55806
Feb 10, 2026
MSE
RMSE
MAE
Updated 4d ago
Evaluation Results
Method
Method
Links
MSE
RMSE
MAE
MILE-RefHumEval
Dimension=Relevance
2026.02
0.495
0.704
0.557
MILE-RefHumEval
Dimension=Consistency
2026.02
0.596
0.772
0.567
MILE-RefHumEval
Dimension=Fluency
2026.02
0.641
0.801
0.685
MILE-RefHumEval
Dimension=Coherence
2026.02
0.927
0.963
0.761
G-Eval-4
Dimension=Relevance
2026.02
0.978
0.989
0.811
G-Eval-4
Dimension=Coherence
2026.02
1.258
1.122
0.91
G-Eval-4
Dimension=Consistency
2026.02
1.849
1.36
1.083
G-Eval-4
Dimension=Fluency
2026.02
4.473
2.115
2.037
Feedback
Search any
task
Search any
task