Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Rethinking ImageNet Pre-training

About

We report competitive results on object detection and instance segmentation on the COCO dataset using standard models trained from random initialization. The results are no worse than their ImageNet pre-training counterparts even when using the hyper-parameters of the baseline system (Mask R-CNN) that were optimized for fine-tuning pre-trained models, with the sole exception of increasing the number of training iterations so the randomly initialized models may converge. Training from random initialization is surprisingly robust; our results hold even when: (i) using only 10% of the training data, (ii) for deeper and wider models, and (iii) for multiple tasks and metrics. Experiments show that ImageNet pre-training speeds up convergence early in training, but does not necessarily provide regularization or improve final target task accuracy. To push the envelope we demonstrate 50.9 AP on COCO object detection without using any external data---a result on par with the top COCO 2017 competition results that used ImageNet pre-training. These observations challenge the conventional wisdom of ImageNet pre-training for dependent tasks and we expect these discoveries will encourage people to rethink the current de facto paradigm of `pre-training and fine-tuning' in computer vision.

Kaiming He, Ross Girshick, Piotr Doll\'ar• 2018

Related benchmarks

TaskDatasetResultRank
Object DetectionCOCO 2017 (val)
AP50.9
2454
Instance SegmentationCOCO 2017 (val)--
1144
Image ClassificationDTD
Accuracy74.72
419
ClassificationCars
Accuracy77.63
314
Image ClassificationGTSRB
Accuracy98.93
291
Image ClassificationCUB
Accuracy72.07
249
Image ClassificationCIFAR10
Accuracy97.45
240
Image ClassificationCaltech101
Accuracy96.11
162
Image ClassificationEuroSAT
Accuracy98.82
83
Image ClassificationFlowers
Accuracy90.25
83
Showing 10 of 17 rows

Other info

Code

Follow for update