Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Layer Collapse Can be Induced by Unstructured Pruning

About

Unstructured pruning is a popular compression method for efficiently reducing model parameters. However, while it effectively decreases the number of parameters, it is commonly believed that unstructured pruning cannot shorten the computational critical path, i.e., the maximum number of layers traversed during forward propagation. In this paper, we study when and how unstructured pruning can yield structural effects. For rectifier-activated networks, we introduce the notion of neuron entropy, which quantifies the degree of nonlinearity utilization. We show that magnitude-based pruning naturally lowers this entropy, sometimes down to zero-entropy layers that become linearizable and can thus be removed. Building on this insight, we propose a method that leverages "unstructured" pruning to favor sparsity in low-entropy layers, enabling their complete removal. We validate the phenomenon across CNNs, Vision Transformers, and NLP models: unstructured pruning can induce effective layer removal with little or no performance degradation in over-parameterized networks.

Zhu Liao, Victor Qu\'etu, Van-Tam Nguyen, Enzo Tartaglione• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet (test)
Top-1 Accuracy66.17
291
Image ClassificationPACS (test)
Average Accuracy95.1
254
Image ClassificationVLCS (test)--
65
Recognizing Textual EntailmentRTE (test)
Accuracy66.06
26
Image ClassificationCIFAR-10 (test)
Top-1 Accuracy93.26
24
Image ClassificationSVIRO (test)
Top-1 Accuracy99.98
21
Sentiment AnalysisSST-2 (test)
Top-1 Accuracy89.79
12
Image ClassificationCIFAR-10 (test)
Top-1 Accuracy89.2
6
Multiple-choice Question AnsweringMMLU High school US history (test)
Top-1 Accuracy36.76
6
Multiple-choice Question AnsweringMMLU High school US history (val)
Top-1 Accuracy38.49
6
Showing 10 of 10 rows

Other info

Follow for update