Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Perceptual Representations for Gaming NR-VQA with Multi-Task FR Signals

About

No-reference video quality assessment (NR-VQA) for gaming videos is challenging due to limited human-rated datasets and unique content characteristics including fast motion, stylized graphics, and compression artifacts. We present MTL-VQA, a multi-task learning framework that uses full-reference metrics as supervisory signals to learn perceptually meaningful features without human labels for pretraining. By jointly optimizing multiple full-reference (FR) objectives with adaptive task weighting, our approach learns shared representations that transfer effectively to NR-VQA. Experiments on gaming video datasets show MTL-VQA achieves performance competitive with state-of-the-art NR-VQA methods across both MOS-supervised and label-efficient/self-supervised settings.

Yu-Chih Chen, Michael Wang, Chieh-Dun Wen, Kai-Siang Ma, Avinab Saha, Li-Heng Chen, Alan Bovik• 2026

Related benchmarks

TaskDatasetResultRank
Video Quality AssessmentLIVE-YT-Gaming
SRCC0.8486
37
Video Quality AssessmentLIVE-Meta MCG
SRCC0.9434
16
Video Quality AssessmentYouTube UGC Gaming 108 videos
SRCC0.8292
15
Showing 3 of 3 rows

Other info

Follow for update