Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning both Weights and Connections for Efficient Neural Networks

About

Neural networks are both computationally intensive and memory intensive, making them difficult to deploy on embedded systems. Also, conventional networks fix the architecture before training starts; as a result, training cannot improve the architecture. To address these limitations, we describe a method to reduce the storage and computation required by neural networks by an order of magnitude without affecting their accuracy by learning only the important connections. Our method prunes redundant connections using a three-step method. First, we train the network to learn which connections are important. Next, we prune the unimportant connections. Finally, we retrain the network to fine tune the weights of the remaining connections. On the ImageNet dataset, our method reduced the number of parameters of AlexNet by a factor of 9x, from 61 million to 6.7 million, without incurring accuracy loss. Similar experiments with VGG-16 found that the number of parameters can be reduced by 13x, from 138 million to 10.3 million, again with no loss of accuracy.

Song Han, Jeff Pool, John Tran, William J. Dally• 2015

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag
Accuracy49.13
1460
Image ClassificationImageNet-1k (val)
Top-1 Accuracy57.2
1453
Language ModelingWikiText-2
Perplexity (PPL)4.98
841
Question AnsweringARC Challenge
Accuracy34.9
749
Image ClassificationFashion MNIST (test)
Accuracy89.2
568
Question AnsweringARC Easy
Accuracy64.02
386
Natural Language InferenceRTE
Accuracy57.04
367
Language ModelingC4
Perplexity9.38
321
Class-conditional Image GenerationImageNet 256x256 (train)
IS70.36
305
Image ClassificationImageNet (test)
Top-1 Accuracy68.38
291
Showing 10 of 56 rows

Other info

Follow for update