Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Rigging the Lottery: Making All Tickets Winners

About

Many applications require sparse neural networks due to space or inference time restrictions. There is a large body of work on training dense networks to yield sparse networks for inference, but this limits the size of the largest trainable sparse model to that of the largest trainable dense model. In this paper we introduce a method to train sparse neural networks with a fixed parameter count and a fixed computational cost throughout training, without sacrificing accuracy relative to existing dense-to-sparse training methods. Our method updates the topology of the sparse network during training by using parameter magnitudes and infrequent gradient calculations. We show that this approach requires fewer floating-point operations (FLOPs) to achieve a given level of accuracy compared to prior techniques. We demonstrate state-of-the-art sparse training results on a variety of networks and datasets, including ResNet-50, MobileNets on Imagenet-2012, and RNNs on WikiText-103. Finally, we provide some insights into why allowing the topology to change during the optimization can overcome local minima encountered when the topology remains static. Code used in our work can be found in github.com/google-research/rigl.

Utku Evci, Trevor Gale, Jacob Menick, Pablo Samuel Castro, Erich Elsen• 2019

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy77.18
3518
Image ClassificationCIFAR-10 (test)
Accuracy94.42
3381
Object DetectionCOCO 2017 (val)--
2454
Image ClassificationImageNet-1K 1.0 (val)
Top-1 Accuracy77.1
1866
Image ClassificationImageNet-1k (val)
Top-1 Accuracy75.1
1453
Image ClassificationImageNet (val)
Top-1 Acc75.1
1206
Instance SegmentationCOCO 2017 (val)--
1144
Image GenerationCIFAR-10 (test)--
471
Image ClassificationImageNet (val)
Accuracy75
300
Image GenerationCIFAR-10
Inception Score8.93
178
Showing 10 of 36 rows

Other info

Code

Follow for update