Transformer for Image Quality Assessment
About
Transformer has become the new standard method in natural language processing (NLP), and it also attracts research interests in computer vision area. In this paper we investigate the application of Transformer in Image Quality (TRIQ) assessment. Following the original Transformer encoder employed in Vision Transformer (ViT), we propose an architecture of using a shallow Transformer encoder on the top of a feature map extracted by convolution neural networks (CNN). Adaptive positional embedding is employed in the Transformer encoder to handle images with arbitrary resolutions. Different settings of Transformer architectures have been investigated on publicly available image quality databases. We have found that the proposed TRIQ architecture achieves outstanding performance. The implementation of TRIQ is published on Github (https://github.com/junyongyou/triq).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Quality Assessment | TID 2013 (test) | Mean SRCC0.846 | 124 | |
| Image Quality Assessment | CSIQ (test) | SRCC0.825 | 103 | |
| Image Quality Assessment | KADID-10k (test) | SRCC0.85 | 91 | |
| No-Reference Image Quality Assessment | KonIQ-10k | SROCC0.892 | 73 | |
| No-Reference Image Quality Assessment | CSIQ | SROCC0.828 | 73 | |
| No-Reference Image Quality Assessment | LIVE | SROCC0.949 | 53 | |
| No-Reference Image Quality Assessment | KADID-10K | PLCC0.855 | 49 | |
| No-Reference Image Quality Assessment | LIVEFB | PLCC0.581 | 42 | |
| No-Reference Image Quality Assessment | TID 2013 | SRCC0.846 | 40 | |
| Face Image Quality Assessment | CGFIQA-40k (test) | PLCC0.9813 | 37 |