Fighting Fake News: Image Splice Detection via Learned Self-Consistency
About
Advances in photo editing and manipulation tools have made it significantly easier to create fake imagery. Learning to detect such manipulations, however, remains a challenging problem due to the lack of sufficient amounts of manipulated training data. In this paper, we propose a learning algorithm for detecting visual image manipulations that is trained only using a large dataset of real photographs. The algorithm uses the automatically recorded photo EXIF metadata as supervisory signal for training a model to determine whether an image is self-consistent -- that is, whether its content could have been produced by a single imaging pipeline. We apply this self-consistency model to the task of detecting and localizing image splices. The proposed method obtains state-of-the-art performance on several image forensics benchmarks, despite never seeing any manipulated images at training. That said, it is merely a step in the long quest for a truly general purpose visual forensics tool.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Forgery Detection | DSO-1 | AUC76.4 | 25 | |
| Image Forgery Detection | Columbia | AUC0.976 | 25 | |
| Image Forgery Detection | Coverage | AUC0.498 | 25 | |
| Image Manipulation Localization | Coverage | -- | 16 | |
| Image Forgery Detection | VIPP | AUC0.617 | 15 | |
| Image Forgery Detection | CocoGlide | AUC52.6 | 15 | |
| Image Forgery Detection | CASIA v1+ | AUC49 | 15 | |
| Image Forgery Detection | NIST16 | AUC0.504 | 15 | |
| Image Manipulation Localization | Columbia | F1 (best)88 | 14 | |
| Image Forgery Localization | DSO-1 | F1 (best)0.577 | 14 |