Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Picking Winning Tickets Before Training by Preserving Gradient Flow

About

Overparameterization has been shown to benefit both the optimization and generalization of neural networks, but large networks are resource hungry at both training and test time. Network pruning can reduce test-time resource requirements, but is typically applied to trained networks and therefore cannot avoid the expensive training process. We aim to prune networks at initialization, thereby saving resources at training time as well. Specifically, we argue that efficient training requires preserving the gradient flow through the network. This leads to a simple but effective pruning criterion we term Gradient Signal Preservation (GraSP). We empirically investigate the effectiveness of the proposed method with extensive experiments on CIFAR-10, CIFAR-100, Tiny-ImageNet and ImageNet, using VGGNet and ResNet architectures. Our method can prune 80% of the weights of a VGG-16 network on ImageNet at initialization, with only a 1.6% drop in top-1 accuracy. Moreover, our method achieves significantly better performance than the baseline at extreme sparsity levels.

Chaoqi Wang, Guodong Zhang, Roger Grosse• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy73.28
3518
Image ClassificationCIFAR-10 (test)
Accuracy93.3
3381
Image ClassificationImageNet (val)--
1206
Image ClassificationImageNet (val)
Accuracy67.21
300
Visual Question AnsweringUltra-MedVQA Task 1
Accuracy37.87
26
Visual Question AnsweringUltra-MedVQA Task 4
Accuracy62.68
26
Visual Question AnsweringUltra-MedVQA Task 3
Accuracy74.83
26
Visual Question AnsweringUltra-MedVQA Task 2
Accuracy76.9
26
Visual Question AnsweringUltra-MedVQA Task 5
Accuracy70.85
26
Visual Question AnsweringUltra-MedVQA Task 6
Accuracy82.54
26
Showing 10 of 18 rows

Other info

Follow for update