Pre-training without Natural Images
About
Is it possible to use convolutional neural networks pre-trained without any natural images to assist natural image understanding? The paper proposes a novel concept, Formula-driven Supervised Learning. We automatically generate image patterns and their category labels by assigning fractals, which are based on a natural law existing in the background knowledge of the real world. Theoretically, the use of automatically generated images instead of natural images in the pre-training phase allows us to generate an infinite scale dataset of labeled images. Although the models pre-trained with the proposed Fractal DataBase (FractalDB), a database without natural images, does not necessarily outperform models pre-trained with human annotated datasets at all settings, we are able to partially surpass the accuracy of ImageNet/Places pre-trained models. The image representation with the proposed FractalDB captures a unique feature in the visualization of convolutional layers and attentions.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-100 | -- | 622 | |
| Image Classification | Stanford Cars | Accuracy86 | 477 | |
| Image Classification | Oxford Flowers 102 | Accuracy98.3 | 172 | |
| Image Classification | CIFAR-10 | Accuracy96.8 | 101 | |
| Image Classification | ImageNet-100 | Accuracy88.3 | 84 | |
| Image Classification | PASCAL VOC 2012 | Accuracy80.6 | 10 | |
| Image Classification | Places30 | Accuracy78.4 | 10 | |
| Image Retrieval | ImageNet-100 (val) | Top-1 Accuracy17.24 | 10 |