Towards Efficient Image Deblurring for Edge Deployment
About
Image deblurring is a critical stage in mobile image signal processing pipelines, where the ability to restore fine structures and textures must be balanced with real-time constraints on edge devices. While recent deep networks such as transformers and activation-free architectures achieve state-of-the-art (SOTA) accuracy, their efficiency is typically measured in FLOPs or parameters, which do not correlate with latency on embedded hardware. We propose a hardware-aware adaptation framework that restructures existing models through sensitivity-guided block substitution, surrogate distillation, and training-free multi-objective search driven by device profiling. Applied to the 36-block NAFNet baseline, the optimized variants achieve up to 55% reduction in GMACs compared to the recent transformer-based SOTA while maintaining competitive accuracy. Most importantly, on-device deployment yields a 1.25X latency improvement over the baseline. Experiments on motion deblurring (GoPro), defocus deblurring (DPDD), and auxiliary benchmarks (RealBlur-J/R, HIDE) demonstrate the generality of the approach, while comparisons with prior efficient baselines confirm its accuracy-efficiency trade-off. These results establish feedback-driven adaptation as a principled strategy for bridging the gap between algorithmic design and deployment-ready deblurring models.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Deblurring | GoPro (test) | PSNR33.76 | 585 | |
| Image Deblurring | RealBlur-J (test) | PSNR26.51 | 226 | |
| Image Deblurring | HIDE (test) | PSNR30.36 | 207 | |
| Deblurring | RealBlur-R (test) | PSNR33.82 | 147 | |
| Defocus Deblurring | DPD Outdoor Scenes | PSNR21.89 | 17 | |
| Defocus De-blurring | DPDD Indoor Scenes | PSNR26.87 | 11 |