Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Snapshot Ensembles: Train 1, get M for free

About

Ensembles of neural networks are known to be much more robust and accurate than individual networks. However, training multiple deep networks for model averaging is computationally expensive. In this paper, we propose a method to obtain the seemingly contradictory goal of ensembling multiple neural networks at no additional training cost. We achieve this goal by training a single neural network, converging to several local minima along its optimization path and saving the model parameters. To obtain repeated rapid convergence, we leverage recent work on cyclic learning rate schedules. The resulting technique, which we refer to as Snapshot Ensembling, is simple, yet surprisingly effective. We show in a series of experiments that our approach is compatible with diverse network architectures and learning tasks. It consistently yields lower error rates than state-of-the-art single models at no additional training cost, and compares favorably with traditional network ensembles. On CIFAR-10 and CIFAR-100 our DenseNet Snapshot Ensembles obtain error rates of 3.4% and 17.4% respectively.

Gao Huang, Yixuan Li, Geoff Pleiss, Zhuang Liu, John E. Hopcroft, Kilian Q. Weinberger• 2017

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy87.11
3518
Image ClassificationCIFAR-10 (test)--
3381
Image ClassificationTiny ImageNet (test)--
265
Image ClassificationImageNet (test)--
235
Image ClassificationSUN397 (test)
Top-1 Accuracy65.79
136
Commonsense ReasoningARC Challenge
Accuracy64.9
132
Commonsense ReasoningARC-E
Accuracy85.2
62
Commonsense ReasoningWG-M
Accuracy73.8
18
Commonsense ReasoningWG-S
Accuracy66.7
18
Commonsense ReasoningBoolQ
Accuracy85.4
18
Showing 10 of 18 rows

Other info

Follow for update