Multi Task Denoiser Training for Solving Linear Inverse Problems
About
Plug-and-Play Priors (PnP) and Regularisation by Denoising (RED) have established that image denoisers can effectively replace traditional regularisers in linear inverse problem solvers for tasks like super-resolution, demosaicing, and inpainting. It is now well established in the literature that a denoiser's residual links to the gradient of the image log prior (Miyasawa and Tweedie), enabling iterative, gradient ascent-based image generation (e.g., diffusion models), as well as new methods for solving inverse problems. Building on this, we propose enhancing Kadkhodaie and Simoncelli's gradient-based inverse solvers by fine-tuning the denoiser within the iterative solving process itself. Training the denoiser end-to-end across the solver framework and simultaneously across multiple tasks yields a single, versatile denoiser optimised for inverse problems. We demonstrate that even a simple baseline model fine-tuned this way achieves an average PSNR improvement of +1.34 dB across six diverse inverse problems while reducing the required iterations. Furthermore, we analyse the fine-tuned denoiser's properties, finding that its optimisation objective implicitly shifts from minimising standard denoising error (MMSE) towards approximating an ideal prior gradient specifically tailored for guiding inverse recovery.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Compressed Sensing | BSDS100 (test) | PSNR27.09 | 78 | |
| Freq. SR | BSD100 (test) | PSNR31.23 | 8 | |
| Inpainting | BSD100 (test) | PSNR37.75 | 8 | |
| Rand. Basis | BSD100 (test) | PSNR31.45 | 8 | |
| SR-x2 | BSD100 (test) | PSNR29.78 | 8 | |
| Demosaicing | BSD100 (test) | PSNR38.22 | 8 |