Our new X account is live! Follow @wizwand_team for updates
Search any
task
Feedback
Search any
task
SOTA Multi-modal Understanding benchmarks and papers with code | Wizwand
Our new X account is live! Follow @wizwand_team for updates
Home
/
Tasks
Multi-modal Understanding
Benchmarks
Dataset Name
SOTA Method
Dataset Name
SOTA Method
Metric
Trend
Results
Last Updated
LLaVA-Bench Wild
GPT4V
LLaVA^W Score
91.2
52
2d ago
MMBench (dev)
Mini-Gemini-HD
Overall Score
80.6
40
2d ago
SEED-Bench (overall)
CSR
Overall Score
62.9
40
3d ago
MMBench EN
Gemini-2.5-Pro
Overall Score
86.3
39
3d ago
MMVet
GPT-4o
Accuracy
76.2
35
3d ago
MMBench EN
InternVL2.5-78B
Accuracy
88.3
34
2d ago
MMBench V1.1
InternVL3.5-38B
Accuracy
87.03
22
3d ago
MMBench
Oryx-1.5
Mean Accuracy
86.3
16
3d ago
SEED-Bench all (val)
LLaVA-Next
Accuracy
65.6
14
3d ago
MMBench (test)
ScalSelect
MMBench Accuracy (En)
65.3
12
3d ago
MM-Vet
LLaVA1.5-BPO
Rec
46.9
11
3d ago
MME v1.0 (test)
LLaVA-1.5-13B
MME^P Score
1,531.3
10
3d ago
MME
Vanilla Attack
M(cl, nc)
90.85
9
3d ago
General Multi-modal Evaluation Suite (VQAv2, GQA, VisWiz, ScienceQA-IMG, TextVQA, POPE, MMBench, MM-Vet) standard (test val)
MoE-LLaVA-Phi DeRS-SM
VQAv2 Accuracy
77.7
9
3d ago
MMBench EN 1.1 (dev)
Pixtral-12B
Accuracy
84.31
8
3d ago
Q-Bench (test)
mPLUG-Owl2
Overall Score
62.9
8
3d ago
CV-Bench, BLINK, RealWorldQA, MathVista, MMStar, MMVet
LATTE
Average Score
53.8
8
3d ago
SEED-IMG
Vanilla 7B
Accuracy
69.7
7
2d ago
SEED-IMG I
Vanilla
Accuracy
66.9
7
2d ago
LLaVA Multi-modal Evaluation Suite (GQA, MMB, MME, POPE, SQA, VQAv2, TextVQA, MMMU, SEED-I) v1.6 (test)
Vanilla 13B
GQA Accuracy
65.4
5
2d ago
MMBench
TaLo
S-I-U
52.5
3
3d ago
LLaVA Multi-modal Evaluation Suite (GQA, MMB, MME, POPE, SQA, VQAv2, TextVQA, MMMU, SEED-I) v1.6 (test val)
-
GQA
-
0
3d ago
Showing 22 of 22 rows
25 / page
50 / page
100 / page
1
Search any
task
Search any
task
Terms of Service
FAQs