Super-resolution of Omnidirectional Images Using Adversarial Learning
About
An omnidirectional image (ODI) enables viewers to look in every direction from a fixed point through a head-mounted display providing an immersive experience compared to that of a standard image. Designing immersive virtual reality systems with ODIs is challenging as they require high resolution content. In this paper, we study super-resolution for ODIs and propose an improved generative adversarial network based model which is optimized to handle the artifacts obtained in the spherical observational space. Specifically, we propose to use a fast PatchGAN discriminator, as it needs fewer parameters and improves the super-resolution at a fine scale. We also explore the generative models with adversarial learning by introducing a spherical-content specific loss function, called 360-SS. To train and test the performance of our proposed model we prepare a dataset of 4500 ODIs. Our results demonstrate the efficacy of the proposed method and identify new challenges in ODI super-resolution for future investigations.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Super-Resolution | ODI-SR (test) | WS-PSNR23.28 | 85 | |
| Super-Resolution | SUN 360 Panorama (test) | WS-PSNR21.48 | 62 | |
| Omnidirectional Image Super-Resolution | ODI-SR | WS-PSNR21.65 | 16 | |
| Omnidirectional Image Super-Resolution | SUN 360 Panorama | WS-PSNR21.48 | 16 |