MDS-VQA: Model-Informed Data Selection for Video Quality Assessment
About
Learning-based video quality assessment (VQA) has advanced rapidly, yet progress is increasingly constrained by a disconnect between model design and dataset curation. Model-centric approaches often iterate on fixed benchmarks, while data-centric efforts collect new human labels without systematically targeting the weaknesses of existing VQA models. Here, we describe MDS-VQA, a model-informed data selection mechanism for curating unlabeled videos that are both difficult for the base VQA model and diverse in content. Difficulty is estimated by a failure predictor trained with a ranking objective, and diversity is measured using deep semantic video features, with a greedy procedure balancing the two under a constrained labeling budget. Experiments across multiple VQA datasets and models demonstrate that MDS-VQA identifies diverse, challenging samples that are particularly informative for active fine-tuning. With only a 5% selected subset per target domain, the fine-tuned model improves mean SRCC from 0.651 to 0.722 and achieves the top gMAD rank, indicating strong adaptation and generalization.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Video Quality Assessment | YouTube-UGC (test) | SRCC0.819 | 36 | |
| Video Quality Assessment | CGVDS (test) | SRCC0.874 | 10 | |
| Video Quality Assessment | LIVE-Livestream (test) | SRCC0.632 | 10 | |
| Video Quality Assessment | YouTube-SFV SDR (test) | SRCC0.731 | 10 | |
| Video Quality Assessment | YouTube-SFV HDR2SDR (test) | SRCC50.7 | 10 | |
| Video Quality Assessment | AIGVQA-DB (test) | SRCC0.769 | 10 | |
| Video Quality Assessment | YouTube-UGC, CGVDS, LIVE-Livestream, YouTube-SFV SDR, YouTube-SFV HDR2SDR, AIGVQA-DB | SRCC1 | 10 |