From Patches to Pictures (PaQ-2-PiQ): Mapping the Perceptual Space of Picture Quality
About
Blind or no-reference (NR) perceptual picture quality prediction is a difficult, unsolved problem of great consequence to the social and streaming media industries that impacts billions of viewers daily. Unfortunately, popular NR prediction models perform poorly on real-world distorted pictures. To advance progress on this problem, we introduce the largest (by far) subjective picture quality database, containing about 40000 real-world distorted pictures and 120000 patches, on which we collected about 4M human judgments of picture quality. Using these picture and patch quality labels, we built deep region-based architectures that learn to produce state-of-the-art global picture quality predictions as well as useful local picture quality maps. Our innovations include picture quality prediction architectures that produce global-to-local inferences as well as local-to-global inferences (via feedback).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Novel View Synthesis | Tanks&Temples (test) | PSNR15.77 | 257 | |
| Novel View Synthesis | Mip-NeRF 360 (test) | PSNR16.3 | 184 | |
| Video Quality Assessment | KoNViD-1k | SROCC0.613 | 183 | |
| Image Quality Assessment | CSIQ | SRC0.899 | 150 | |
| Image Quality Assessment | TID 2013 (test) | Mean SRCC0.488 | 134 | |
| No-Reference Image Quality Assessment | CSIQ | SROCC0.899 | 121 | |
| Image Quality Assessment | KonIQ | SRCC0.872 | 119 | |
| Blind Image Quality Assessment | FLIVE | SRCC0.571 | 115 | |
| No-Reference Image Quality Assessment | KADID-10K | SROCC0.84 | 115 | |
| No-Reference Image Quality Assessment | KonIQ-10k | SROCC0.872 | 111 |