Fully Trainable and Interpretable Non-Local Sparse Models for Image Restoration
About
Non-local self-similarity and sparsity principles have proven to be powerful priors for natural image modeling. We propose a novel differentiable relaxation of joint sparsity that exploits both principles and leads to a general framework for image restoration which is (1) trainable end to end, (2) fully interpretable, and (3) much more compact than competing deep learning architectures. We apply this approach to denoising, jpeg deblocking, and demosaicking, and show that, with as few as 100K parameters, its performance on several standard benchmarks is on par or better than state-of-the-art methods that may have an order of magnitude or more parameters.
Bruno Lecouat, Jean Ponce, Julien Mairal• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Denoising | BSD68 grayscale (test) | PSNR37.95 | 101 | |
| Color Denoising | CBSD68 (test) | PSNR (dB)36.4 | 62 | |
| Grayscale Image Denoising | BDS68 (test) | PSNR31.7 | 35 | |
| Grayscale Image Denoising | Urban100 (test) | PSNR32.72 | 34 | |
| Blind Image Denoising | CBSD68 | PSNR40.43 | 30 | |
| Grayscale Image Denoising | Set12 (test) | -- | 29 | |
| Color Denoising | CBSD68 | Average PSNR (σ=50)28.05 | 8 | |
| JPEG artifact reduction | Classic5 qf=10 (test) | PSNR29.61 | 7 | |
| JPEG artifact reduction | Classic5 qf=20 (test) | PSNR31.78 | 7 | |
| JPEG artifact reduction | Classic5 qf=30 (test) | PSNR33.06 | 7 |
Showing 10 of 14 rows