Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Resolution-robust Large Mask Inpainting with Fourier Convolutions

About

Modern image inpainting systems, despite the significant progress, often struggle with large missing areas, complex geometric structures, and high-resolution images. We find that one of the main reasons for that is the lack of an effective receptive field in both the inpainting network and the loss function. To alleviate this issue, we propose a new method called large mask inpainting (LaMa). LaMa is based on i) a new inpainting network architecture that uses fast Fourier convolutions (FFCs), which have the image-wide receptive field; ii) a high receptive field perceptual loss; iii) large training masks, which unlocks the potential of the first two components. Our inpainting network improves the state-of-the-art across a range of datasets and achieves excellent performance even in challenging scenarios, e.g. completion of periodic structures. Our model generalizes surprisingly well to resolutions that are higher than those seen at train time, and achieves this at lower parameter&time costs than the competitive baselines. The code is available at \url{https://github.com/saic-mdal/lama}.

Roman Suvorov, Elizaveta Logacheva, Anton Mashikhin, Anastasia Remizova, Arsenii Ashukha, Aleksei Silvestrov, Naejin Kong, Harshith Goka, Kiwoong Park, Victor Lempitsky• 2021

Related benchmarks

TaskDatasetResultRank
Image InpaintingPlaces2 (test)
PSNR22.694
68
InpaintingImageNet
LPIPS0.061
54
Image InpaintingCelebA-HQ
LPIPS0.028
42
Image InpaintingFFHQ (test)
FID23.2
40
InpaintingCelebA-HQ
LPIPS0.028
36
InpaintingPlaces2 Wide Mask 512x512 (test)
FID1.72
30
Image InpaintingFFHQ 256x256 (val)
FID32.7035
22
Image InpaintingPlaces2 512x512 (test)
LPIPS0.086
20
InpaintingPlaces narrow masks 512 x 512
FID0.63
20
InpaintingPlaces wide masks 512 x 512
FID2.21
20
Showing 10 of 75 rows
...

Other info

Code

Follow for update