Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Adaptive Image Quality Assessment via Teaching Large Multimodal Model to Compare

About

While recent advancements in large multimodal models (LMMs) have significantly improved their abilities in image quality assessment (IQA) relying on absolute quality rating, how to transfer reliable relative quality comparison outputs to continuous perceptual quality scores remains largely unexplored. To address this gap, we introduce Compare2Score-an all-around LMM-based no-reference IQA (NR-IQA) model, which is capable of producing qualitatively comparative responses and effectively translating these discrete comparative levels into a continuous quality score. Specifically, during training, we present to generate scaled-up comparative instructions by comparing images from the same IQA dataset, allowing for more flexible integration of diverse IQA datasets. Utilizing the established large-scale training corpus, we develop a human-like visual quality comparator. During inference, moving beyond binary choices, we propose a soft comparison method that calculates the likelihood of the test image being preferred over multiple predefined anchor images. The quality score is further optimized by maximum a posteriori estimation with the resulting probability matrix. Extensive experiments on nine IQA datasets validate that the Compare2Score effectively bridges text-defined comparative levels during training with converted single image quality score for inference, surpassing state-of-the-art IQA models across diverse scenarios. Moreover, we verify that the probability-matrix-based inference conversion not only improves the rating accuracy of Compare2Score but also zero-shot general-purpose LMMs, suggesting its intrinsic effectiveness.

Hanwei Zhu, Haoning Wu, Yixuan Li, Zicheng Zhang, Baoliang Chen, Lingyu Zhu, Yuming Fang, Guangtao Zhai, Weisi Lin, Shiqi Wang• 2024

Related benchmarks

TaskDatasetResultRank
Image Quality AssessmentSPAQ
SRCC0.86
191
Image Quality AssessmentCSIQ
SRC0.705
138
Image Quality AssessmentAGIQA-3K
SRCC0.671
112
Image Quality AssessmentCSIQ (test)
SRCC0.705
103
Image Quality AssessmentKADID
SRCC45.3
95
Image Quality AssessmentPIPAL
SRCC34.2
95
Blind Image Quality AssessmentFLIVE
SRCC0.413
86
Image Quality AssessmentKonIQ
SRCC0.91
82
Image Quality AssessmentSPAQ (test)
SRCC0.86
77
No-Reference Image Quality AssessmentKADID (test)
SROCC0.453
42
Showing 10 of 16 rows

Other info

Follow for update