Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Exploring CLIP for Assessing the Look and Feel of Images

About

Measuring the perception of visual content is a long-standing problem in computer vision. Many mathematical models have been developed to evaluate the look or quality of an image. Despite the effectiveness of such tools in quantifying degradations such as noise and blurriness levels, such quantification is loosely coupled with human language. When it comes to more abstract perception about the feel of visual content, existing methods can only rely on supervised models that are explicitly trained with labeled data collected via laborious user study. In this paper, we go beyond the conventional paradigms by exploring the rich visual language prior encapsulated in Contrastive Language-Image Pre-training (CLIP) models for assessing both the quality perception (look) and abstract perception (feel) of images in a zero-shot manner. In particular, we discuss effective prompt designs and show an effective prompt pairing strategy to harness the prior. We also provide extensive experiments on controlled datasets and Image Quality Assessment (IQA) benchmarks. Our results show that CLIP captures meaningful priors that generalize well to different perceptual assessments. Code is avaliable at https://github.com/IceClear/CLIP-IQA.

Jianyi Wang, Kelvin C.K. Chan, Chen Change Loy• 2022

Related benchmarks

TaskDatasetResultRank
Image Quality AssessmentSPAQ
SRCC0.901
250
Image Quality AssessmentCSIQ
SRC0.862
150
Image Quality AssessmentAGIQA-3K
SRCC0.844
131
Image Quality AssessmentKADID
SRCC65.4
128
Image Quality AssessmentKonIQ-10k
SRCC0.895
126
Image Quality AssessmentPIPAL
SRCC43.1
123
No-Reference Image Quality AssessmentCSIQ
SROCC0.89
121
Image Quality AssessmentKonIQ
SRCC0.905
119
Blind Image Quality AssessmentFLIVE
SRCC0.602
115
No-Reference Image Quality AssessmentKonIQ-10k
SROCC0.895
111
Showing 10 of 123 rows
...

Other info

Code

Follow for update